Most active commenters
  • akhayam(6)
  • onesphere(3)

←back to thread

1901 points l2silver | 13 comments | | HN request time: 1.57s | source | bottom

Maybe you've created your own AR program for wearables that shows the definition of a word when you highlight it IRL, or you've built a personal calendar app for your family to display on a monitor in the kitchen. Whatever it is, I'd love to hear it.
1. akhayam ◴[] No.35730472[source]
About 8 years back, I was leading an engineering team which was the escalation path for customer support. We were sitting on a large corpus of support tickets but didn't have any insights. I was amazed when word2vec came out and blew my mind. So I built a language model that trained on support ticket data. I modeled system logs attached to support tickets as an NLP model to predict what was going to fail next and for which customer.

Never made it to prod but was a great tool for me to see where I want to budget my team's time.

This is way before all the LLM and Generator models, but it was such a fun project.

replies(3): >>35739376 #>>35739939 #>>35756307 #
2. UweSchmidt ◴[] No.35739376[source]
>This is way before all the LLM and Generator models, but it was such a fun project.

That means something more sophisticated has to exist today and should be commercially available. Can anyone explain to what extent companies use this stuff in their interaction with customers, and how successful is it? (Somehow I still see AI still as one-off things people do for fun or AI being used to hype up rather mundane software.)

replies(1): >>35741792 #
3. onesphere ◴[] No.35739939[source]
We have a corpus or database of programs that follow logic but with no simulation, so it represents knowledge to solve a problem yet all we have control over is the parameters (inputs). In this case, input is functional logical content (a program), describing the resolution of corpus details. The model solves its integrated, corporate logic, and our output is an interpretation of that individual program.

Now our task is to swap out this entire database for something like it, but not exactly the same. The output becomes the input to this new matrix. The individual program persists, but everything is the next generation. With a little book-keeping, the programs do our will...

replies(1): >>35741733 #
4. akhayam ◴[] No.35741733[source]
Don't think I quite follow. Is the new program (operating on the output of the earlier program) supposed to reason about why you are seeing the result that you are seeing? Or is it doing more post processing to make the earlier output directly consumable by your corporate systems.
replies(1): >>35743143 #
5. akhayam ◴[] No.35741792[source]
It should exist and would be super powerful considering all the recent advancements in language ML. Here was the mental model of my model: the canonical representation (i.e. a representation after taking out run-time populated fields) of a log line represents the smallest meaningful unit of this "log language": _a word_. Taking this analogy further, an event is a collection of logs that occur together (mostly in order)—just like words spoken together form _a sentence_. Finally, collections of events that occur in close proximity (in time) represent _paragraphs_, while paragraphs occurring in a certain order constitute _chapters_. Using this mental model opens the door to apply all the new AI techniques for text extraction, summarization and generation to extract the semantic structure of any "log language" and then learn and classify behaviors observed at run-time. The eventual objective function is not generation though--it's reasoning with the optimal FP-TP tradeoff on a ROC curve.

I haven't seen anyone do it yet. Maybe companies like Splunk and Elastic will take a lead here. I am happy to engage, advise and contribute if there is an open source project around this. Has anyone else seen something remotely close to this?

replies(1): >>35742889 #
6. UweSchmidt ◴[] No.35742889{3}[source]
Very interesting, thanks!
7. onesphere ◴[] No.35743143{3}[source]
The new program’s purpose could be to do more post processing to make the interpretation of that earlier program directly consumable (inter-generationally), or it could simply start producing more problems to solve.
replies(1): >>35744881 #
8. akhayam ◴[] No.35744881{4}[source]
Gotcha! That makes sense. I would recommend looking at LangChain though, as it does a good job at modeling multi-stage learning / inference environments.
replies(1): >>35751965 #
9. onesphere ◴[] No.35751965{5}[source]
Integration API: https://thetaplane.com/ai/langchain/api

Inspired by: https://github.com/daveebbelaar/langchain-experiments/blob/m...

replies(1): >>35759223 #
10. mgsouth ◴[] No.35756307[source]
There's a YC startup that's doing something along these lines. They recently posted something, don't remember if it was on front page or an "YC startup jobs" email.
replies(1): >>35763070 #
11. akhayam ◴[] No.35759223{6}[source]
Wow... that was quick. Would love to see what results you get.
12. fakedang ◴[] No.35763070[source]
Mendable.ai?
replies(1): >>35765367 #
13. akhayam ◴[] No.35765367{3}[source]
Nah... Mendable seems to be searching in documentation. This idea is to extract an execution model of every component in your system from the logs of that component. We can think of the prediction step as a complete separate machine model for now.

Anything else comes to mind?