Show HN: FinetuneDB – AI fine-tuning platform to create custom LLMs

finetunedb.com

152 points by felix089 5 days ago

Hey HN! We’re building FinetuneDB (https://finetunedb.com/), an LLM fine-tuning platform. It enables teams to easily create and manage high-quality datasets, and streamlines the entire workflow from fine-tuning to serving and evaluating models with domain experts. You can check out our docs here: (https://docs.finetunedb.com/)

FinetuneDB exists because creating and managing high-quality datasets is a real bottleneck when fine-tuning LLMs. The quality of your data directly impacts the performance of your fine-tuned models, and existing tools didn’t offer an easy way for teams to build, organize, and iterate on their datasets. We’ve been working closely with our pilot customers, both AI startups and more traditional businesses like a large newspaper, which is fine-tuning models on their articles to automate content generation in their tone of voice.

The platform is built with an end-to-end workflow in mind, from dataset building, fine-tuning, serving, and evaluating outputs. The centerpiece is a version-controlled, no-code dataset manager where you can upload existing datasets in JSONL, use production data, or collaborate with domain experts to create high-quality datasets for custom use cases. We also offer evaluation workflows that allow non-technical contributors to annotate data, review model outputs, and refine responses (LLM-as-judge also available).

We offer:

- A free tier for developers and hobbyists who want to streamline dataset management.

- Business-tier with full feature access for teams, using per-seat pricing.

- A custom tier for model hosting, custom integrations, and self-hosting.

Most users still use OpenAI models, but if you're working with open-source LLMs, we offer pay-as-you-go pricing for serverless inference for Llama and Mistral models with up to €100 in free credits to get started.

We're in public beta right now, so any feedback—whether it’s about features, usability, or anything else—would be incredibly valuable.

If you've worked on fine-tuning models before or are curious about custom LLMs, we’d love to hear from you. Our goal is to make the fine-tuning process more accessible and help more companies leverage their data and domain experts to create custom LLMs.

Thanks for checking it out!

luke-stanley 5 days ago

This sounds interesting to me! I will check it out in more detail. I saw 3.5-turbo mentioned which is more expensive and as I understand it, it's usually less good as a base model, if I see 3.5-turbo before I see 4o-mini, and don't see 4o-mini, I might wonder if things are behind! I hope that's fair feedback for a quick reaction. I have done a bunch of fine-tuning before, locally and with 4o-mini, and there's often a lot of time spent suboptimally on wrangling data so I'm interested in this category of product for sure, if it helps more than it costs me.

  • felix089 5 days ago

    You can fine-tune any OpenAI model that is available in your account, so both 3.5-turbo and 4o-mini work, but mini replaced 3.5 for most use cases. Where did you see 3.5 mentioned over mini?

    • JulianWasTaken 5 days ago

      It's in at least 3 places on the homepage -- in 2 screenshots and in the first main code block you show. This was the first thing I noticed too, so I think it's good feedback from OP.

      • farouqaldori 5 days ago

        Yes, it's outdated. Thanks for the feedback. 4o-mini is the new king for sure!

      • felix089 5 days ago

        Ah I see, yes agreed, we'll update that asap! Thanks

trentontri 5 days ago

Why bury the pricing information under the documentation? The problem with these platforms is that it is unclear how much bandwidth/money your use case will require to actually train and run a successful LLM.

The world needs products like this that are local first and open source. Enable me train an open source LLM on my M2 Macbook with a desktop app then I'll consider giving you my money. App developers integrating LLM's need to be able to experiment and see the potential before storing everything on the cloud.

  • farouqaldori 5 days ago

    We are working on a dedicated pricing page with all relevant information. Pricing in docs is just temporary. With that being said, new users get free credits to try out the platform without spending anything.

    We've built the platform primarily for companies that serve LLMs in production, so even if we allowed you to fine-tune on device, sooner or later you will find yourself in a position where you want to deploy the model.

    We want to streamline this whole process, end-to-end.

    With that being said, I do agree that we shouldn't store everything on the cloud, this is what we're doing about it:

    1. Any data in FinetuneDB like evals, logs, datasets etc. can be exported or deleted.

    2. Fine-tuned model weights for OS models can be downloaded.

    3. Using our inference stack is not a requirement. Many users are happy with only the dataset manager (which is 100% free).

    4. We are exploring options to integrate external databases and storage providers with FinetuneDB, allowing datasets to be stored off our servers

farouqaldori 5 days ago

Hey all, co-founder here happy to answer any questions!

  • krishnasangeeth 5 days ago

    Nice product overall. I had some feedback and questions

    feedback :

    1. Since you guys have support for multiple models, it would be cleaner and more correct to give the API some name which doesn't start with openAI.

    2. sdk using other languages like Python in `show code` would be nice.

    3. It was a bit confusing to figure out how to fine tune the model, would be nice if it was explicitly available as a side pane.

    Questions:

    1. Can you speak a bit about your tech stack if that's alright

    2. How do you currently scale inference if there is more incoming requests coming in?

    • farouqaldori 5 days ago

      Thank you so much!

      1. Where exactly did you see this? There are internal FinetuneDB API keys, and external API keys like OpenAI. Though it's confusing, I agree!

      2. Work in progress.

      3. I agree, thanks for the feedback.

      There are multiple components working together, so it's hard to define a single tech stack. When it comes to the web app, Remix is my framework of choice and can highly recommend it.

      • FunkyFreddy 5 days ago

        Congrats on the launch, UI looks sleek! Is tracking logs available in the free plan?

        • felix089 5 days ago

          Thanks, and yes, tracking logs is included in the free plan!

  • namanyayg 5 days ago

    What benefits does this bring me vs just using OpenAI's official tools?

    • felix089 5 days ago

      Other co-founder here, so we offer more specific features around iterating on your datasets and include domain experts in this workflow. And I'd argue that you also want your datasets not necessarily with your foundation model provider like OpenAI, so you have the option to test with and potentially switch to open-source models.

  • skerit 5 days ago

    Is it possible to fine-tune language models using plain text completions, or is it necessary to use datasets consisting of structured conversations?

    • felix089 5 days ago

      Yes, you can fine-tune using plain text completions. You don't need structured conversations unless you want conversational abilities. Plain text works great if you want the model to generate text in a specific style or domain. It all depends on what you're trying to achieve.

      • skerit 5 days ago

        Nice.

        And about the cost of finetuning: is there a difference in price when only training the model on completions?

        • felix089 5 days ago

          The cost depends on the number of tokens processed, so fine-tuning on completions costs the same per token as any other data.

  • rmbyrro 5 days ago

    What's the cost of fine tuning and then serving a model, say Llama 3 8B or 70B? I couldn't find anything on the website...

    • felix089 5 days ago

      Hi, current pricing for Llama 3.1 8B for example is: Training Tokens: $2 / 1M, Input and Output Tokens: $0.30 / 1M. We'll update pricing on the website shortly to reflect this.

perhapsAnLLM 5 days ago

Awesome work! Really clean UI - who are your competitors that offer a similar "end-to-end workflow" UI for LLMs? I'm typically in a Jupyter notebook for this type of thing but a neat and snappy web app could certainly help streamline some workflows.

  • farouqaldori 5 days ago

    Thanks for the feedback! More than happy to learn more about your workflow if you'd like to share (farouq@finetunedb.com)

  • felix089 5 days ago

    Happy to hear you like the UI, ease of use is key for us. Would love for you to give it a try, any feedback welcome!

ilovefood 5 days ago

Looks pretty cool, congrats so far! Do you allow downloading the fine tuned model for local inference?

  • felix089 5 days ago

    Thank you, and yes that is possible. Which model are you looking to fine-tune?

    • ilovefood 5 days ago

      If that's the case then I'll try the platform out :) I want to finetune Codestral or Qwen2.5-coder on a custom codebase. Thank you for the response! Are there some docs or infos about the compatibility of the downloaded models, meaning will they work right away with llama.cpp?

      • farouqaldori 5 days ago

        We don't support Codestral or Qwen2.5-coder right out of the box for now, but depending on your use-case we certainly could add it.

        We utilize LoRA for smaller models, and qLoRA (quantized) for 70b+ models to improve training speeds, so when downloading model weights, what you get is the weights & adapter_config.json. Should work with llama.cpp!

inSenCite 5 days ago

Wow this is really cool, congrats on the launch!

Does the platform also help speed up the labelling of semi-structured data? I have a use case where I need to take data in word, ppt, pdf; label paragraphs / sections which could then be used to fine tune a model

  • felix089 5 days ago

    Thank you! We currently don't support direct labeling, but if you can extract the text, our platform helps you organize it for fine-tuning. What use case are you looking to train the model for?

    • inSenCite 5 days ago

      Ah, ok. the text extraction is manageable enough, I will carve out a smaller subset so I can give your platform a go. The use case is professional services contract creation and redlining based on reference documents.

      • felix089 5 days ago

        Okay thanks for sharing, I think that's the way to go with the subset. Feel free to reach out if you need anything, more than happy to take a closer look!

fpgaminer 4 days ago

I gave it a try, but when I tried to start a finetune of Llama 3.1 8B it just gave an error every time. I also encountered several server errors just navigating to different pages.

  • felix089 4 days ago

    Thanks for giving it a try, and sorry to hear you're having issues, could you please share the errors you received either here or via founders@finetunedb.com? Many users successfully fine-tuned models today, so it would be great to learn what the specific problem is. Thanks!

KaoruAoiShiho 5 days ago

Am I able to upload a book and have it respond truthfully to the book in a way that's superior to NotebookLM or similar? Generally most long context solutions are very poor. Or does the data have to be in a specific format?

  • felix089 5 days ago

    To get the outcome you want, RAG (retrieval augmented generation) would be the way to go, not fine-tuning. Fine-tuning doesn't make the model memorize specific content like a book. It teaches new behaviors or styles. RAG allows the model to access and reference the book during inference. Our platform focuses on fine-tuning with structured datasets, so data needs to be in a specific format.

    This is a very common topic, so I wrote a blog post that explains the difference between fine-tuning and RAG if you're interested: https://finetunedb.com/blog/fine-tuning-vs-rag

    • thomashop 5 days ago

      These days, I'd say the easiest and most effective approach is to put the whole book in the context of one of the longer context models.

      • felix089 5 days ago

        Agreed, for this use case probably the easiest way to go.

        • swyx 5 days ago

          (and most expensive)

      • KaoruAoiShiho 5 days ago

        Not really, for something like gemini the accuracy and performance is very poor.

        • farouqaldori 5 days ago

          The magic behind NotebookLM can't be replicated only with fine-tuning. It's all about the workflow, from the chunking strategy, to retrieval etc.

          For a defined specific use-case it's certainly possible to beat their performance, but things get harder when you try to create a general solution.

          To answer your question, the format of the data depends entirely on the use-case and how many examples you have. The more examples you have, the more flexible you can be.

    • hodanli 5 days ago

      I dont think it is a big deal but you can use your own image or give credit to openai presentation on YouTube.

monkeydust 5 days ago

If I wanted to tune a LLama model on say MermaidJS or PlantUML to improve performance beyond what they can do today with this product be a good fit?

  • farouqaldori 5 days ago

    Yes for sure. But your mileage may vary, and it all depends on the quality of the dataset that you build up.

    Happy to discuss this in detail, how do you measure performance?

cl42 5 days ago

Was looking for a solution like this for a few weeks, and started coding my own yesterday. Thank you for launching! Excited to give it a shot.

Question: when do you expect to release your Python SDK?

  • farouqaldori 5 days ago

    There hasn't been a significant demand for the Python SDK yet, so for now we suggest interacting with the API directly.

    With that being said, feel free to email us with your use-case, I could build the SDK within a few days!

    • rmbyrro 5 days ago

      If you currently have an SDK in any of the 5 major languages, or if your API is well documented in a structured way, it should be very easy to write ab SDK in Python, Go, anything LLMs know well.

    • cl42 5 days ago

      Main requirement is to programmatically send my chat logs. Not a big deal though, thanks!

      • farouqaldori 5 days ago

        Ah I see, got it. For now the API should work fine for that!

  • felix089 5 days ago

    Very happy to hear, please do reach out to us with any feedback or questions via founders@finetunedb.com

EliBullockPapa 5 days ago

Awesome but pricing seems a little high right now, and you're missing Gemini Flash, the very cheapest fine tunable model that I know of.

  • farouqaldori 5 days ago

    Which part of the pricing seems high, platform or token pricing? Both?

    About Gemini Flash, we add new model providers entirely based on feedback. Gemini is next on the roadmap!

    • kouteiheika 5 days ago

      > Which part of the pricing seems high, platform or token pricing? Both?

      You said that you do only LoRA finetuning and your pricing for Llama 3.1 8B is $2/1M tokens. To me this does seem high. I can do full finetuning (so not just a LoRA!) of Llama 3.1 8B for something like ~$0.2/M if I rent a 4090 on RunPod, and ~$0.1/M if I just get the cheapest 4090 I can find on the net.

      • farouqaldori 4 days ago

        That's true when looking solely at fine-tuning costs. In theory, you could fine-tune a model locally and only cover electricity expenses. However, we provide a complete end-to-end workflow that simplifies the entire process.

        Once a model is fine-tuned, you can run inference on Llama 3.2 3B for as low as $0.12 per million tokens. This includes access to logging, evaluation, and continuous dataset improvement through collaboration, all without needing to set up GPUs or manage the surrounding infrastructure yourself.

        Our primary goal is to provide the best dataset for your specific use case. If you decide to deploy elsewhere to reduce costs, you always have the option to download the model weights.

        • kouteiheika 4 days ago

          Sure, I'm just comparing the baseline costs of finetuning. Assuming you own the hardware and optimize the training I'm guessing you could easily get the costs significantly lower than $0.1/M tokens (considering I can get the $0.1/M right now using publicly rented GPUs, and whoever I'm renting the GPU from is still making money on me), and if you're only doing LoRA that cost would go down even further (don't have the numbers on hand because I never do LoRA finetuning, so I have no idea how much faster that is per token compared to full finetuning).

          So your $2/M tokens for LoRA finetuning tells me that you either have a very (per dollar) inefficient finetuning pipeline (e.g. renting expensive GPUs from AWS) and need such a high price to make any money, or that you're charging ~20x~30x more than it costs you. If it's the latter - fair enough, some people will pay a premium for all of the extra features! If it's the former - you might want to consider optimizing your pipeline to bring those costs down. (:

Aditya_Garg 5 days ago

How do I get the 100 dollars in credit? Did a preliminary fine tune, but would like to rigorously test out your tech for my article.

  • felix089 5 days ago

    Thanks for giving it a try! The 100 is for the pro plan, by default you should have 10 in your account, but happy to add more, please email me with your account email, and I'll top it up: founders@finetunedb.com - Thanks!

I_am_tiberius 5 days ago

Looks nice. What is the price and what does it depend on?

  • felix089 5 days ago

    Thanks! We have a free tier with limited features. Our pro plan starts at €50 per seat per month and includes all features. Teams often collaborate with domain experts to create datasets. And for custom integrations, we offer custom plans on request.

    More details here: https://docs.finetunedb.com/getting-started/pricing

    Any specific features or use cases you're interested in?

    • martypitt 5 days ago

      Congrats on the launch - definitely interested.

      Some minor feedback - I went to the website to look for pricing (scanned the header bar), and couldn't find it.

      Didn't think to look in the docs, as it's almost always available from the homepage.

      Appreciate you linking it here,but if I hadn't come from HN, I'd assume this is a "contact us for pricing" situation, which is a bit of a turnoff.

      • felix089 5 days ago

        Nice to hear, and agreed thanks for the feedback, our pricing page will be up shortly. If any other questions come up RE product, please reach out!

jpnagel 5 days ago

looks nice, have been looking into fine-tuning for a while, will check it out!

fabbe199913 4 days ago

Very interesting!

  • felix089 4 days ago

    Happy to hear, any questions let us know!

groa 5 days ago

[dead]