"supervised learning: model is trained on the labeled data
unsupervised learning: no labels provided - model learns by finding patterns , structure and groupings in the data.
Semi-supervised learning: use small set of labels to guide learning for the larger pool of unlabeled data.
reinforcement learning: leans by interacting with students the environment, receives reward and penalties based on actions
self supervised: no labelled data . The model makes its own practice problems by"
Anchal V. - "supervised learning: model is trained on the labeled data
unsupervised learning: no labels provided - model learns by finding patterns , structure and groupings in the data.
Semi-supervised learning: use small set of labels to guide learning for the larger pool of unlabeled data.
reinforcement learning: leans by interacting with students the environment, receives reward and penalties based on actions
self supervised: no labelled data . The model makes its own practice problems by"See full answer
"Designing an evaluation framework for ads ranking is crucial for optimizing the effectiveness and relevance of ads displayed to users. Here's a comprehensive framework that you can use:
Define Objectives and Key Performance Indicators (KPIs):**
\\Click-Through Rate (CTR):\\ The ratio of clicks to impressions, indicating the effectiveness of an ad in attracting user attention.
\\Conversion Rate:\\ The ratio of conversions (e.g., sign-ups, purchases) to clicks, measuring how well"
Ajay P. - "Designing an evaluation framework for ads ranking is crucial for optimizing the effectiveness and relevance of ads displayed to users. Here's a comprehensive framework that you can use:
Define Objectives and Key Performance Indicators (KPIs):**
\\Click-Through Rate (CTR):\\ The ratio of clicks to impressions, indicating the effectiveness of an ad in attracting user attention.
\\Conversion Rate:\\ The ratio of conversions (e.g., sign-ups, purchases) to clicks, measuring how well"See full answer
"There can be multiple effects on adjusting the context window of LLM, some I can think of are below:
If window size is large then more tokens are in context which could increase memory and compute costs because of O(n2) attention complexity.
Larger window can help in better responses in multi turn conversations but attention dilution can also happen."
Raja raghudeep E. - "There can be multiple effects on adjusting the context window of LLM, some I can think of are below:
If window size is large then more tokens are in context which could increase memory and compute costs because of O(n2) attention complexity.
Larger window can help in better responses in multi turn conversations but attention dilution can also happen."See full answer
"C : Okay. So I would want to start with knowing what is the product for which we have to build a recommendation system.
I : This is a photo sharing product.
C : Okay. So is this something on the lines of Instagram?
I : Yes
C : Okay. And are we a new product co or we have some current product built already?
I : You can assume yourself.
C : Okay. Is there any demography or country we are targeting?
I : No, this is a global product
C : Okay. So, the biggest goal of any product recommendation system"
Kartikeya N. - "C : Okay. So I would want to start with knowing what is the product for which we have to build a recommendation system.
I : This is a photo sharing product.
C : Okay. So is this something on the lines of Instagram?
I : Yes
C : Okay. And are we a new product co or we have some current product built already?
I : You can assume yourself.
C : Okay. Is there any demography or country we are targeting?
I : No, this is a global product
C : Okay. So, the biggest goal of any product recommendation system"See full answer
"The interviewer hinted that a two-tower recommender system might be a suitable approach, using user history to embed users and pages separately and train on view or interaction data.
Instead, I proposed a different approach that I felt was more aligned with how knowledge is structured in Confluence:
I designed a system using a graph database to model the relationships between Confluence pages. Each page is a node, and edges represent content-based references. For example, when one article"
Clayton P. - "The interviewer hinted that a two-tower recommender system might be a suitable approach, using user history to embed users and pages separately and train on view or interaction data.
Instead, I proposed a different approach that I felt was more aligned with how knowledge is structured in Confluence:
I designed a system using a graph database to model the relationships between Confluence pages. Each page is a node, and edges represent content-based references. For example, when one article"See full answer
"Overall a very good summary of ranking problems.
I wish there are more details in the post on depth of modeling. One can talk more about advance ML architectures like DCN, BertRec etc"
Kapil D. - "Overall a very good summary of ranking problems.
I wish there are more details in the post on depth of modeling. One can talk more about advance ML architectures like DCN, BertRec etc"See full answer
" Thanks a lot for showing us how a recommender system can be build. I see it was proposed to use Collaborative filtering which is user - item matrix having dimension N * M (where N - number os users and M - number of songs). Though, it was explained how it gonna be built, it is still unclear how all users and songs features are going to be used. In that matrix we have values in cell (lets say i, j) like 1 - a specific user (i) clicked on song (j) when it was recommended or it is 0 when the user"
Dinar M. - " Thanks a lot for showing us how a recommender system can be build. I see it was proposed to use Collaborative filtering which is user - item matrix having dimension N * M (where N - number os users and M - number of songs). Though, it was explained how it gonna be built, it is still unclear how all users and songs features are going to be used. In that matrix we have values in cell (lets say i, j) like 1 - a specific user (i) clicked on song (j) when it was recommended or it is 0 when the user"See full answer
"At a high level, the core challenge here revolves around building an effective recommendation algorithm for news.
News is an inherently diverse category, spanning various topics and catering to a wide array of user types and personas, such as adults, business professionals, general readers, or specific cohorts with unique interests. Consequently, developing a single, one-size-fits-all recommendation algorithm is not feasible.
To enhance the personalization of the news recommendation algorithm,"
Sai vuppalapati M. - "At a high level, the core challenge here revolves around building an effective recommendation algorithm for news.
News is an inherently diverse category, spanning various topics and catering to a wide array of user types and personas, such as adults, business professionals, general readers, or specific cohorts with unique interests. Consequently, developing a single, one-size-fits-all recommendation algorithm is not feasible.
To enhance the personalization of the news recommendation algorithm,"See full answer
"Gradient Descent is an optimisation strategy used in several supervised learning models. It is the technique for finding the optimum solution of an objective function. Typically, for a linear regression use case, it is used to find the weights and bias that produce the lowest loss.
It involves computing the partial derivative of the objective function with respect to the weight and bias vectors. To find the optima of the function, the derivative is equated to 0, and iteratively the weight and b"
Megha V. - "Gradient Descent is an optimisation strategy used in several supervised learning models. It is the technique for finding the optimum solution of an objective function. Typically, for a linear regression use case, it is used to find the weights and bias that produce the lowest loss.
It involves computing the partial derivative of the objective function with respect to the weight and bias vectors. To find the optima of the function, the derivative is equated to 0, and iteratively the weight and b"See full answer
"DNNs can learn hierarchical features, with each layer learning progressively more abstract features, and generalizes better. SNNs are better for simplier problems involving smaller datasets and if low latency is required."
Louie Z. - "DNNs can learn hierarchical features, with each layer learning progressively more abstract features, and generalizes better. SNNs are better for simplier problems involving smaller datasets and if low latency is required."See full answer