"Constrain the requirements (The final product should be somewhat like Neptune AI and also enable for A/B testing)
Do some back of the envelope calculations
Start explaning your system with appropriate architecture diagrams
Optimize for scale
Answer any questions
Wrap-up"
Varun G. - "Constrain the requirements (The final product should be somewhat like Neptune AI and also enable for A/B testing)
Do some back of the envelope calculations
Start explaning your system with appropriate architecture diagrams
Optimize for scale
Answer any questions
Wrap-up"See full answer
"There are 2 main methods
Intrinsic Evaluation
i) Preplexity
ii) BLEU
Extrinsic Evaluation
i) Response consistency/ Correctness / Factual score/ Security
However, this question requires a follow-up question and clarification about where we are going to use the LLM models."
Mayank M. - "There are 2 main methods
Intrinsic Evaluation
i) Preplexity
ii) BLEU
Extrinsic Evaluation
i) Response consistency/ Correctness / Factual score/ Security
However, this question requires a follow-up question and clarification about where we are going to use the LLM models."See full answer
"Designing an evaluation framework for ads ranking is crucial for optimizing the effectiveness and relevance of ads displayed to users. Here's a comprehensive framework that you can use:
Define Objectives and Key Performance Indicators (KPIs):**
\\Click-Through Rate (CTR):\\ The ratio of clicks to impressions, indicating the effectiveness of an ad in attracting user attention.
\\Conversion Rate:\\ The ratio of conversions (e.g., sign-ups, purchases) to clicks, measuring how well"
Ajay P. - "Designing an evaluation framework for ads ranking is crucial for optimizing the effectiveness and relevance of ads displayed to users. Here's a comprehensive framework that you can use:
Define Objectives and Key Performance Indicators (KPIs):**
\\Click-Through Rate (CTR):\\ The ratio of clicks to impressions, indicating the effectiveness of an ad in attracting user attention.
\\Conversion Rate:\\ The ratio of conversions (e.g., sign-ups, purchases) to clicks, measuring how well"See full answer
"Since question asks about pipeline. I assume the question is about metrics across many dimensions not just prediction Model performance.
For the ML Model: I can use accuracy, precision, recall, F1 if it is classification model. In case it is regression model RMSE is good metric for many problems.
Data: ML system needs good quality data. The system has to track missing data rate. Distribution of features, if there is no drift from original feature distributions during the training.
Pipeline h"
Alex N. - "Since question asks about pipeline. I assume the question is about metrics across many dimensions not just prediction Model performance.
For the ML Model: I can use accuracy, precision, recall, F1 if it is classification model. In case it is regression model RMSE is good metric for many problems.
Data: ML system needs good quality data. The system has to track missing data rate. Distribution of features, if there is no drift from original feature distributions during the training.
Pipeline h"See full answer
Machine Learning Engineer
Machine Learning
+1 more
🧠 Want an expert answer to a question? Saving questions lets us know what content to make next.
"The adjusting context window size in LLM change trade off between reasoning capability, accuracy, computation cost. It influence how attention mechanist allocate resources across the input.
Longer context window let it you input greater number of words and have more context to generate proper next token. However llms have lost in the middle issue. They remember the beginning of text and end of text but lost information located in the middle of long input. Another problem is Attention Dilution."
Alex N. - "The adjusting context window size in LLM change trade off between reasoning capability, accuracy, computation cost. It influence how attention mechanist allocate resources across the input.
Longer context window let it you input greater number of words and have more context to generate proper next token. However llms have lost in the middle issue. They remember the beginning of text and end of text but lost information located in the middle of long input. Another problem is Attention Dilution."See full answer
"supervised learning: model is trained on the labeled data
unsupervised learning: no labels provided - model learns by finding patterns , structure and groupings in the data.
Semi-supervised learning: use small set of labels to guide learning for the larger pool of unlabeled data.
reinforcement learning: leans by interacting with students the environment, receives reward and penalties based on actions
self supervised: no labelled data . The model makes its own practice problems by"
Anchal V. - "supervised learning: model is trained on the labeled data
unsupervised learning: no labels provided - model learns by finding patterns , structure and groupings in the data.
Semi-supervised learning: use small set of labels to guide learning for the larger pool of unlabeled data.
reinforcement learning: leans by interacting with students the environment, receives reward and penalties based on actions
self supervised: no labelled data . The model makes its own practice problems by"See full answer
"C : Okay. So I would want to start with knowing what is the product for which we have to build a recommendation system.
I : This is a photo sharing product.
C : Okay. So is this something on the lines of Instagram?
I : Yes
C : Okay. And are we a new product co or we have some current product built already?
I : You can assume yourself.
C : Okay. Is there any demography or country we are targeting?
I : No, this is a global product
C : Okay. So, the biggest goal of any product recommendation system"
Kartikeya N. - "C : Okay. So I would want to start with knowing what is the product for which we have to build a recommendation system.
I : This is a photo sharing product.
C : Okay. So is this something on the lines of Instagram?
I : Yes
C : Okay. And are we a new product co or we have some current product built already?
I : You can assume yourself.
C : Okay. Is there any demography or country we are targeting?
I : No, this is a global product
C : Okay. So, the biggest goal of any product recommendation system"See full answer
"Hallucinations are evaluated by measuring how often generated outputs contain information that is not supported by trusted sources.
what hallucination means in context:
Intrinsic hallucination: contradicts provided context
Extrinsic hallucination: introduces unsupported facts
Fabrication: confidently incorrect answers"
Hardik saurabh G. - "Hallucinations are evaluated by measuring how often generated outputs contain information that is not supported by trusted sources.
what hallucination means in context:
Intrinsic hallucination: contradicts provided context
Extrinsic hallucination: introduces unsupported facts
Fabrication: confidently incorrect answers"See full answer
"At a high level, the core challenge here revolves around building an effective recommendation algorithm for news.
News is an inherently diverse category, spanning various topics and catering to a wide array of user types and personas, such as adults, business professionals, general readers, or specific cohorts with unique interests. Consequently, developing a single, one-size-fits-all recommendation algorithm is not feasible.
To enhance the personalization of the news recommendation algorithm,"
Sai vuppalapati M. - "At a high level, the core challenge here revolves around building an effective recommendation algorithm for news.
News is an inherently diverse category, spanning various topics and catering to a wide array of user types and personas, such as adults, business professionals, general readers, or specific cohorts with unique interests. Consequently, developing a single, one-size-fits-all recommendation algorithm is not feasible.
To enhance the personalization of the news recommendation algorithm,"See full answer
"Gradient Descent is an optimisation strategy used in several supervised learning models. It is the technique for finding the optimum solution of an objective function. Typically, for a linear regression use case, it is used to find the weights and bias that produce the lowest loss.
It involves computing the partial derivative of the objective function with respect to the weight and bias vectors. To find the optima of the function, the derivative is equated to 0, and iteratively the weight and b"
Megha V. - "Gradient Descent is an optimisation strategy used in several supervised learning models. It is the technique for finding the optimum solution of an objective function. Typically, for a linear regression use case, it is used to find the weights and bias that produce the lowest loss.
It involves computing the partial derivative of the objective function with respect to the weight and bias vectors. To find the optima of the function, the derivative is equated to 0, and iteratively the weight and b"See full answer
"The interviewer hinted that a two-tower recommender system might be a suitable approach, using user history to embed users and pages separately and train on view or interaction data.
Instead, I proposed a different approach that I felt was more aligned with how knowledge is structured in Confluence:
I designed a system using a graph database to model the relationships between Confluence pages. Each page is a node, and edges represent content-based references. For example, when one article"
Clayton P. - "The interviewer hinted that a two-tower recommender system might be a suitable approach, using user history to embed users and pages separately and train on view or interaction data.
Instead, I proposed a different approach that I felt was more aligned with how knowledge is structured in Confluence:
I designed a system using a graph database to model the relationships between Confluence pages. Each page is a node, and edges represent content-based references. For example, when one article"See full answer
"DNNs can learn hierarchical features, with each layer learning progressively more abstract features, and generalizes better. SNNs are better for simplier problems involving smaller datasets and if low latency is required."
Louie Z. - "DNNs can learn hierarchical features, with each layer learning progressively more abstract features, and generalizes better. SNNs are better for simplier problems involving smaller datasets and if low latency is required."See full answer