Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
from flask import Flask, request, jsonify, send_from_directory
|
2 |
from flask_limiter import Limiter
|
3 |
from flask_limiter.util import get_remote_address
|
4 |
-
|
5 |
from PIL import Image
|
6 |
from io import BytesIO
|
7 |
from prodiapy import Prodia
|
@@ -255,7 +255,7 @@ async def generate_image():
|
|
255 |
return jsonify({"status": "error", "error": "Internal Server Error"}), 500
|
256 |
|
257 |
|
258 |
-
configure(api_key="AIzaSyCKc8gUTrb7KjhAwY6NwATaMWIDr5etKFk")
|
259 |
|
260 |
# Set up the model
|
261 |
generation_config = {
|
@@ -284,14 +284,13 @@ safety_settings = [
|
|
284 |
},
|
285 |
]
|
286 |
|
287 |
-
model = GenerativeModel(
|
288 |
model_name="gemini-1.0-pro-001",
|
289 |
generation_config=generation_config,
|
290 |
safety_settings=safety_settings
|
291 |
)
|
292 |
|
293 |
@app.route('/gemini', methods=['GET'])
|
294 |
-
@limiter.limit("30 per minute")
|
295 |
def gemini():
|
296 |
prompt = request.args.get('prompt')
|
297 |
|
@@ -299,6 +298,7 @@ def gemini():
|
|
299 |
return jsonify({'error': 'Prompt parameter is required'}), 400
|
300 |
|
301 |
convo = model.start_chat(history=[prompt])
|
|
|
302 |
response = convo.last.text
|
303 |
|
304 |
return jsonify({'response': response})
|
|
|
1 |
from flask import Flask, request, jsonify, send_from_directory
|
2 |
from flask_limiter import Limiter
|
3 |
from flask_limiter.util import get_remote_address
|
4 |
+
import google.generativeai as genai
|
5 |
from PIL import Image
|
6 |
from io import BytesIO
|
7 |
from prodiapy import Prodia
|
|
|
255 |
return jsonify({"status": "error", "error": "Internal Server Error"}), 500
|
256 |
|
257 |
|
258 |
+
genai.configure(api_key="AIzaSyCKc8gUTrb7KjhAwY6NwATaMWIDr5etKFk")
|
259 |
|
260 |
# Set up the model
|
261 |
generation_config = {
|
|
|
284 |
},
|
285 |
]
|
286 |
|
287 |
+
model = genai.GenerativeModel(
|
288 |
model_name="gemini-1.0-pro-001",
|
289 |
generation_config=generation_config,
|
290 |
safety_settings=safety_settings
|
291 |
)
|
292 |
|
293 |
@app.route('/gemini', methods=['GET'])
|
|
|
294 |
def gemini():
|
295 |
prompt = request.args.get('prompt')
|
296 |
|
|
|
298 |
return jsonify({'error': 'Prompt parameter is required'}), 400
|
299 |
|
300 |
convo = model.start_chat(history=[prompt])
|
301 |
+
convo.send_message(prompt) # Simulating user input with the prompt itself
|
302 |
response = convo.last.text
|
303 |
|
304 |
return jsonify({'response': response})
|