RakeshUtekar commited on
Commit
322d226
·
verified ·
1 Parent(s): 28e78a6

Upload 2 files

Browse files
Files changed (2) hide show
  1. blip_model.py +18 -0
  2. gpt_model.py +30 -0
blip_model.py ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from PIL import Image
2
+ from transformers import BlipForConditionalGeneration, BlipProcessor
3
+
4
+ processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
5
+ model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
6
+
7
+ def extract_image_details(image):
8
+ inputs = processor(images=image, return_tensors="pt")
9
+
10
+ generated_ids = model.generate(
11
+ pixel_values=inputs["pixel_values"],
12
+ max_length=50,
13
+ num_beams=5,
14
+ do_sample=False
15
+ )
16
+ generated_text = processor.batch_decode(generated_ids, skip_special_tokens=True)[0]
17
+ print(f"BLIP Model Description: {generated_text}") # Debugging print statement
18
+ return generated_text
gpt_model.py ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import openai
2
+
3
+
4
+ def generate_response(retrieved_texts, query, max_tokens=150):
5
+ """
6
+ Generates a response based on the retrieved texts and query.
7
+
8
+ Args:
9
+ retrieved_texts (list): List of retrieved text strings.
10
+ query (str): Query string.
11
+ max_tokens (int): Maximum number of tokens for the response.
12
+
13
+ Returns:
14
+ str: Generated response.
15
+ """
16
+ context = "\n".join(retrieved_texts)
17
+ prompt = f"This is the detail about the image: {context}\n\nQuestion: {query}\n\nAnswer:"
18
+
19
+ response = openai.ChatCompletion.create(
20
+ model="gpt-3.5-turbo",
21
+ messages=[
22
+ {"role": "system", "content": "You are a helpful assistant."},
23
+ {"role": "user", "content": prompt}
24
+ ],
25
+ max_tokens=max_tokens,
26
+ n=1,
27
+ stop=None,
28
+ temperature=0.5,
29
+ )
30
+ return response.choices[0].message['content']