apparently working version for LLAMA
This commit is contained in:
parent
e54d16c7a4
commit
00470d0438
2 changed files with 92 additions and 20 deletions
|
@ -1,4 +1,10 @@
|
|||
use base64::{Engine as _, engine::general_purpose::STANDARD};
|
||||
use serde::Deserialize;
|
||||
use serde::Serialize;
|
||||
use serde_json::json;
|
||||
use std::time::Duration;
|
||||
|
||||
|
||||
|
||||
|
||||
// module to hold all code for generating/fetching image descriptions
|
||||
|
@ -21,6 +27,16 @@ pub struct FileConfig {
|
|||
pub caption_extension: String,
|
||||
}
|
||||
|
||||
#[derive(Serialize, Deserialize)]
|
||||
struct LlamaModel {
|
||||
model: String,
|
||||
prompt: String,
|
||||
stream: bool,
|
||||
format: String,
|
||||
suffix: String,
|
||||
images: Vec<String>,
|
||||
keep_alive: i8
|
||||
}
|
||||
// fetch the imagedescription from a file named like the Image
|
||||
pub fn get_description_from_file(image_name: String , file_config: FileConfig) -> Result<String, Box<dyn super::Error>> {
|
||||
//read image caption from a local file that
|
||||
|
@ -74,27 +90,82 @@ pub fn get_description_from_chatgpt(image_name: String, chatgpt_config: self::C
|
|||
}))
|
||||
.send();
|
||||
|
||||
// Improved error handling for API response
|
||||
//if !response.unwrap().status().is_success() {
|
||||
// let error_text = response.unwrap_err();
|
||||
// return Err(format!("OpenAI API error: ", std::error.box(error_text));
|
||||
//}
|
||||
// Improved error handling for API response
|
||||
//if !response.unwrap().status().is_success() {
|
||||
// let error_text = response.unwrap_err();
|
||||
// return Err(format!("OpenAI API error: ", std::error.box(error_text));
|
||||
//}
|
||||
|
||||
let result: super::Value = response.unwrap().json()?;
|
||||
let result: super::Value = response.unwrap().json()?;
|
||||
|
||||
// More detailed error handling for JSON parsing
|
||||
let description = result["choices"]
|
||||
.get(0)
|
||||
.ok_or("No choices in response")?
|
||||
["message"]["content"]
|
||||
.as_str()
|
||||
.ok_or("Invalid content format in response")?
|
||||
.to_string();
|
||||
// More detailed error handling for JSON parsing
|
||||
let description = result["choices"]
|
||||
.get(0)
|
||||
.ok_or("No choices in response")?
|
||||
["message"]["content"]
|
||||
.as_str()
|
||||
.ok_or("Invalid content format in response")?
|
||||
.to_string();
|
||||
|
||||
Ok(description)
|
||||
Ok(description)
|
||||
}
|
||||
|
||||
// fetch images description from own OLLAMA server
|
||||
pub fn get_description_from_ollama(image_name: String, ollama_config: OllamaConfig) -> Result<String, Box<dyn super::Error>> {
|
||||
Ok("Not implemented yet".to_string())
|
||||
// Read and encode image
|
||||
let image_data = std::fs::read(image_name)?;
|
||||
// Base64 encode the image for ChatGTP API
|
||||
let base64_image = STANDARD.encode(image_data);
|
||||
|
||||
|
||||
// Create the JSON payload
|
||||
let payload = json!({
|
||||
"model": ollama_config.ollama_model.to_string(),
|
||||
"prompt": "Can you complete this sentence: This picture shows... Try to be as accurate as possible but keep the description simple and shorter than 5000 characters.",
|
||||
"stream": false,
|
||||
"images": [base64_image]
|
||||
});
|
||||
|
||||
|
||||
// println!("JSON output:\n{}", json.clone());
|
||||
// Create ChatGPT API request
|
||||
// let client = reqwest::blocking::Client::new();
|
||||
let client = reqwest::blocking::ClientBuilder::new()
|
||||
.connect_timeout(Duration::new(30, 0))
|
||||
.timeout(Duration::new(300,0))
|
||||
.connection_verbose(true).build()?;
|
||||
|
||||
|
||||
let response = client
|
||||
.post(ollama_config.ollama_api_url)
|
||||
.header("Content-Type", "application/json")
|
||||
.json(&payload).send();
|
||||
|
||||
// Improved error handling for API response
|
||||
// Check for HTTP errors
|
||||
if response.as_ref().unwrap().status().is_success() {
|
||||
println!("success!");
|
||||
} else if response.as_ref().unwrap().status().is_server_error() {
|
||||
println!("server error!");
|
||||
} else {
|
||||
println!("Something else happened. Status: {:?}", response.as_ref().unwrap().status());
|
||||
}
|
||||
|
||||
|
||||
// Extract response text
|
||||
let result: super::Value = response.unwrap().json()?;
|
||||
let description: String;
|
||||
if !result["response"].is_null() {
|
||||
description = result["response"].to_string();
|
||||
} else if !result["error"].is_null() {
|
||||
description = result["error"].to_string();
|
||||
} else {
|
||||
description = "Could not find response or error from OLLAMA".to_string();
|
||||
}
|
||||
|
||||
|
||||
println!("Description generated by OLLAMA: {}", description.clone());
|
||||
|
||||
Ok(description)
|
||||
|
||||
}
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
use reqwest::{self, multipart};
|
||||
use reqwest::{self};
|
||||
use std::error::Error;
|
||||
|
||||
struct PixelfedConfig {
|
||||
|
@ -59,6 +59,7 @@ pub fn bulk_upload_images(config: &super::Config, images: &[String], batch_num:
|
|||
};
|
||||
println!("Fetching image description from OLLAMA for {}", image_path.to_string());
|
||||
description = super::image_description::get_description_from_ollama(image_path.to_string(), im_config)?;
|
||||
println!("Description generated by OLLAMA: {}", description.clone());
|
||||
media_descriptions.push(description.clone());
|
||||
},
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue