@@ -157,8 +157,15 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
157157
158158func ConvertEmbeddingRequest (request model.GeneralOpenAIRequest ) * EmbeddingRequest {
159159 return & EmbeddingRequest {
160- Model : request .Model ,
161- Prompt : strings .Join (request .ParseInput (), " " ),
160+ Model : request .Model ,
161+ Input : request .ParseInput (),
162+ Options : & Options {
163+ Seed : int (request .Seed ),
164+ Temperature : request .Temperature ,
165+ TopP : request .TopP ,
166+ FrequencyPenalty : request .FrequencyPenalty ,
167+ PresencePenalty : request .PresencePenalty ,
168+ },
162169 }
163170}
164171
@@ -201,15 +208,17 @@ func embeddingResponseOllama2OpenAI(response *EmbeddingResponse) *openai.Embeddi
201208 openAIEmbeddingResponse := openai.EmbeddingResponse {
202209 Object : "list" ,
203210 Data : make ([]openai.EmbeddingResponseItem , 0 , 1 ),
204- Model : "text-embedding-v1" ,
211+ Model : response . Model ,
205212 Usage : model.Usage {TotalTokens : 0 },
206213 }
207214
208- openAIEmbeddingResponse .Data = append (openAIEmbeddingResponse .Data , openai.EmbeddingResponseItem {
209- Object : `embedding` ,
210- Index : 0 ,
211- Embedding : response .Embedding ,
212- })
215+ for i , embedding := range response .Embeddings {
216+ openAIEmbeddingResponse .Data = append (openAIEmbeddingResponse .Data , openai.EmbeddingResponseItem {
217+ Object : `embedding` ,
218+ Index : i ,
219+ Embedding : embedding ,
220+ })
221+ }
213222 return & openAIEmbeddingResponse
214223}
215224
0 commit comments