I have a master file which contains certain text- let's say-
file contains x
the image is of x type
the user is admin
the address is x
and then there 200 other text files containing texts like-
file contains xyz
the image if of abc type
the user is admin
the address if pqrs
I need to match these files up. The result will be true if the files contains the text exactly as is in the master file, with x being different for each file i.e. 'x' in the master can be anything in the other files and the result will be true.What I have come up with is
arr=master.split('\n')
for file in files:
a=[]
file1=file.split('\n')
i=0
for line in arr:
line_list=line.split()
indx=line_list.index('x')
line_list1=line_list[:indx]+line_list[indx+1:]
st1=' '.join(line_list1)
file1_list=file1[i].split()
file1_list1=file1_list[:indx]+file1_list[indx+1:]
st2=' '.join(file1_list1)
if st1!=st2:
a.append(line)
i+=1
which is highly inefficient. Is there a way that I can sort of map the files with master file and generate the differences in some other file?
Is that "universal" unique on the line? For instance, if the key is, indeed,
x
, are you guaranteed thatx
appears nowhere else in the line? Or could the master file have something likeIf you do have a unique key ...
For each line, split the master file on your key
x
. This gives you two pieces for the line, front and back. Then merely check whether the linestartswith
the front part andendswith
the back part. Something likeSee documentation
UPDATE PER OP COMMENT
So long as
x
is unique within the line, you can easily adapt this. As you mention in your comment, you want something like