When they mention 3D Models, are 3D-Video/Picture generating models or 3D object (like Blender) generator models meant? If anyone has some links laying around, both would be interesting use case for me.
I think that one of the most incredible datasets anyone could make would be for a Polars Dataframe library (the extremely efficient training dataset by converting some of the SQL or Pandas datasets.
Data processing is such a huge part of the AI process and depending on how you look at it, extremely expensive or a huge opportunity to reduce costs in both compute and time. The performance improvements that Polars brings to data preparation are simply incredible.
However, since the library is still relatively new and evolving, it's really poorly understood by nearly all of the models, especially building performant custom expressions. I would happily chip in to a project that built a large training dataset that can help us fine-tune efficient data processing LLMs.
grady_vuckovic@reddit
How much of that contains redundant data?
mycall@reddit
How much of this is redundant information?
CheatCodesOfLife@reddit
Thanks for the reminder, I've got to clean up my (private) datasets and half-finished models lol.
PraxisOG@reddit
How much of that contains redundant data?
Qual_@reddit
yes
Blizado@reddit
Happy searching. ðŸ«
I want to have a sci-fi space dataset.
shing3232@reddit
LLM written Star trek story with long term memory:)
Blizado@reddit
For that I would make a extra finetune on top of it. :D
CMD_Shield@reddit
When they mention 3D Models, are 3D-Video/Picture generating models or 3D object (like Blender) generator models meant? If anyone has some links laying around, both would be interesting use case for me.
Thireus@reddit
🫡 Doing my part https://huggingface.co/Thireus/collections 💪
ActivitySpare9399@reddit
I think that one of the most incredible datasets anyone could make would be for a Polars Dataframe library (the extremely efficient training dataset by converting some of the SQL or Pandas datasets.
Data processing is such a huge part of the AI process and depending on how you look at it, extremely expensive or a huge opportunity to reduce costs in both compute and time. The performance improvements that Polars brings to data preparation are simply incredible.
However, since the library is still relatively new and evolving, it's really poorly understood by nearly all of the models, especially building performant custom expressions. I would happily chip in to a project that built a large training dataset that can help us fine-tune efficient data processing LLMs.