| 
					
				 | 
			
			
				@@ -19,9 +19,7 @@ def replace_target(target,seq): 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				             seq[i],seq[i+1],seq[i+2] = -100,-100,-100 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				     return seq 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				 def tokenize_dialogs(dialogs, images, processor): 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				-    # If vocab size is above 128000, use the chat template to generate the tokens as it is from Llama 3 family models 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				     text_prompt = processor.apply_chat_template(dialogs) 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				-    #print("text_prompt",text_prompt) 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				     batch = processor(images=images, text=text_prompt,padding = True, return_tensors="pt") 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				     label_list = [] 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				     for i in range(len(batch["input_ids"])): 
			 |