๐Ÿ‘‹ Connect with me on LinkedIn!

LinkedIn

Manuel Caccone - Actuarial Data Scientist & Open Source Educator

Let's discuss actuarial science, AI, and open source projects!


Model Thumbnail

๐Ÿค– Gemma-3 ActuaryEnough2: Bringing Actuarial AI to Everyone


๐Ÿšฉ Model Description

Gemma-3-actuaryEnough2 is a fine-tuned Gemma-3-270M model trained on over 11,000 actuarial question-answer pairs, purpose-built to translate simple insurance queries into rigorous actuarial technical language. It powers ActuaryEnough and is released as open source for educational and research use.


โœจ Key Features

  • ๐ŸŽฏ Domain-specific: Focused exclusively on actuarial and insurance Q&A.
  • ๐Ÿ“š Educational: Makes complex actuarial terminology accessible for all users.
  • ๐Ÿš€ Efficient: Fine-tuned with Unsloth for rapid, scalable training.
  • ๐Ÿ”“ Open Source: Apache 2.0 License; easy to reuse, adapt, remix.
  • ๐ŸŒ Widget & Demo: Integrated as a live demo on ActuaryEnough.

๐Ÿ’ก Intended Use Cases

  • Education: For students and actuaries in training, or for professionals retraining in actuarial language.
  • Translation: Make practical insurance questions understandable at professional actuarial level.
  • Research: Support for actuarial research, Q&A, and domain adaptation.

Examples

# Premium Calculation Example
Input: "How much should I pay for car insurance? Rephrase:"
Output: "This relates to premium calculation considering risk factors such as exposure units, loss frequency, severity distributions, and loading factors for expenses and profit margins."

๐Ÿ“‚ Training Data

  • Primary Dataset: actuarial-qa-11k - Over 11,000 manually curated actuarial questionโ€“answer pairs
  • Specialized Dataset: actuary-enough-qa-dataset - Actuarial question simplification examples
  • Topics: Life and non-life insurance, risk assessment, regulation, reserves, actuarial mathematics, terminology simplification
  • Language: English
  • Format: Instruction-following format optimized for text generation tasks

๐Ÿ“Š Training Statistics

Metric Value / Range Notes
Epochs ~51 Reached at end of training
Global Steps >68,000
Initial Train Loss ~2.2 At start
Final Train Loss ~1.4 At end
Learning Rate 8e-7 โ†’ โ‰ˆ0 Linear decay throughout training
Gradient Norm 5 โ€“ 15 Generally stable with rare spikes
Hardware RTX 3090, 16-core CPU 24GB VRAM, 94GB RAM, CUDA 12.8, Linux 6.1

๐Ÿ› ๏ธ Dependencies

Python 3.12.11
transformers
torch
unsloth
wandb==0.21.1
pydantic==2.11.7
# ...for full list, check requirements.txt

โš ๏ธ Limitations & Ethics

  • No pricing or decision support: For education and inspiration only, not for real insurance contracts.
  • Not a substitute for an actuary: Always consult professionals for real-world decisions.
  • Coverage: Designed and tested specifically for the insurance/actuarial domain.
  • Training data bias: Outputs may reflect source content.

๐Ÿ’ป Usage Example

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("manuelcaccone/gemma-3-actuaryEnough2")
model = AutoModelForCausalLM.from_pretrained("manuelcaccone/gemma-3-actuaryEnough2")

prompt = "Which factors determine life insurance premiums?"
toks = tokenizer(prompt, return_tensors="pt")
output = model.generate(**toks, max_new_tokens=256, temperature=0.7)
print(tokenizer.decode(output[0], skip_special_tokens=True))

๐ŸŒŸ Related Datasets

This model is part of the ActuaryEnough ecosystem and uses multiple specialized datasets:


๐Ÿ‘ค Author & Citation

@model{caccone2025actuaryenough,
  title={Gemma-3 ActuaryEnough2: A Fine-tuned Model for Actuarial Education},
  author={Caccone, Manuel},
  year={2025},
  publisher={Hugging Face},
  url={https://huggingface.co/manuelcaccone/gemma-3-actuaryEnough2},
  note={Educational model for actuarial science and insurance terminology}
}

๐Ÿ“œ License

Apache 2.0 License โ€” use, modify, and cite for ethical, research, and educational purposes.


๐Ÿค Want to collaborate or discuss actuarial AI?

LinkedIn


Part of the ActuaryEnough open-source education initiativeโ€”bringing actuarial science closer to everyone!

Downloads last month
5
Safetensors
Model size
0.3B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for manuelcaccone/gemma-3-actuaryEnough2

Finetuned
(379)
this model

Dataset used to train manuelcaccone/gemma-3-actuaryEnough2