--- license: cc-by-nc-2.0 language: - cs base_model: - fav-kky/FERNET-C5 --- This is fav-kky/FERNET-C5, fine-tuned with the **Cross-Encoder** architecture on the Czech News Dataset for Semantic Textual Similarity and DaReCzech. The Cross-Encoder architecture processes both input text pieces simultaneously, enabling better accuracy. The model can be used both for Semantic Textual Similarity and re-ranking. **Semantic Textual Similarity**: The model takes two input sentences and evaluates how similar their meanings are. ```python from sentence_transformers import CrossEncoder model = CrossEncoder('ctu-aic/CE-fernet-c5-sfle512', max_length=512) scores = model.predict([["sentence_1", "sentence_2"]]) print(scores) ``` **Re-ranking task**: Given a query, the model assesses all potential passages and ranks them in descending order of relevance. ```python from sentence_transformers import CrossEncoder model = CrossEncoder('ctu-aic/CE-fernet-c5-sfle512', max_length=512) query = "Example query for." documents = [ "Example document one.", "Example document two.", "Example document three." ] top_k = 3 return_documents = True results = model.rank( query=query, documents=documents, top_k=top_k, return_documents=return_documents ) for i, res in enumerate(results): print(f"{i+1}. {res['text']}") ```