AttentionSmithy code co-pilot? #22
Replies: 1 comment
-
I love this idea, thank you for sharing @csmova! This example is really close, but not quite ready - the learned embedding in particular can work, but takes the place of tokenized embeddings, which I don't think is compatible with the end goal of the model. I'm going to work on having a few basic examples (likely simple encoder-only, decoder-only, and encoder-decoder examples) that could be used as reference for something like GitInjest so it can automate creating models like this. That should provide enough breadth of examples that an LLM should be able to implement a given model. Then making a code co-pilot for any interested parties would be a cinch. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I really like this package and am excited to use it! I'm more of an analytical chemist and am not too familiar with the technical coding details of implementing a transformer. Having this 'lego block' approach to building the models seems useful. I've seen transformers being used a lot for proteomics data, and I have a bit of intuition for the machine learning side of such applications, but actually building a transformer seems daunting.
I wanted to just quickly see what the simplest possible proteomics transformer might look like using AttentionSmithy, before I dove deeper into your repo. For example, would I still be needing to code up multiple .py files and hundreds of lines of code? Or would it just be a couple dozen?
Your code is really well structured, so I was thinking maybe ChatGPT could act as 'interactive documentation', where I could chat with it about how to implement your package. I used GitInjest to upload your repo to ChatGPT and had it suggest a basic proteomics transformer model for me. I'm posting as this might be helpful for others. You can view my full chat here
I also uploaded the .txt file from GitInjest, which is attached here, but I'm not sure how much this helped since ChatGPT had the github link anyway.
xomicsdatascience-attentionsmithy.txt
As you should be able to see in the chat link, I asked ChatGPT for some starter code for an example proteomics project. It seemed to grasp your project structure well. Here is the output, can you tell if these a sensible starting point of using AttentionSmithy?:
Is it really this simple (assuming I had real data formatted correctly)? If so, building a custom GPT specifically for AttentionSmity might be worth it (think like a code co-pilot, but tailored with docuemntation and instructions to make AttentionSmithy implementations and codebase Q&A really simple). If this seems promising, perhaps we could work together to do this.
Happy to hear thoughts.
Beta Was this translation helpful? Give feedback.
All reactions