this is the placeholder model card for the (finess-benchmark-space)[https://huggingface.co/spaces/enzoescipy/finesse-benchmark-space] and its (database)[https://huggingface.co/datasets/enzoescipy/finesse-benchmark-results].

import torch
from typing import List
from transformers import AutoConfig, PreTrainedModel  # Optional: for loading configs

from finesse_benchmark.interfaces import  FinesseSynthesizer

# --- Custom Embedder Example ---
# Uncomment and customize this class for your embedder.

class AverageSynthesizer(FinesseSynthesizer):
    """
    Average Synthesizer: Computes the mean of input embeddings without using any model.
    """
    def __init__(self, config_path: str):
        super().__init__()
        # No model to load for average pooling
        print(f"{self.__class__.__name__} initialized - Average pooling ready.")

    def synthesize(self, embeddings: torch.Tensor, **kwargs) -> torch.Tensor:
        """
        Average synthesis: Compute the mean along the sequence dimension.
        
        Args:
            embeddings: torch.Tensor of shape (batch, seq_len, embedding_dim)
            **kwargs: Additional arguments
        
        Returns:
            torch.Tensor of shape (batch, embedding_dim)
        """
        return embeddings.mean(dim=1)

    def device(self):
        return "cpu"

it just averages the embedding vectors.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support