"I would conduct a sample z-test because we have enough samples and the population variance is known.
H1: average monthly spending per user is $50
H0: average monthly spending per user is greater $50
One-sample z-test
x_bar = $85
mu = $50
s = $20
n = 100
x_bar - mu / (s / sqrt(n) = 17.5
17.5 is the z-score that we will need to associate with its corresponding p-value. However, the z-score is very high,
so the p-value will be very close to zero, which is much less than the standa"
Lucas G. - "I would conduct a sample z-test because we have enough samples and the population variance is known.
H1: average monthly spending per user is $50
H0: average monthly spending per user is greater $50
One-sample z-test
x_bar = $85
mu = $50
s = $20
n = 100
x_bar - mu / (s / sqrt(n) = 17.5
17.5 is the z-score that we will need to associate with its corresponding p-value. However, the z-score is very high,
so the p-value will be very close to zero, which is much less than the standa"See full answer
"Clarification question: How many subscription plans are offered by Tinder ?
If there is more than one subscription plan, then we need to ask is the fluctuation happening across all plans or in a particular one ?
Assumption: Let's say lower priced subscription plan is showing the most fluctuation and there are only two types of plans
In this subscription plan which age group is showing the most fluctuation (18-24,25-30, 30+ etc) ?
Is there any seasonality trend observed (eg: placemen"
Srijita P. - "Clarification question: How many subscription plans are offered by Tinder ?
If there is more than one subscription plan, then we need to ask is the fluctuation happening across all plans or in a particular one ?
Assumption: Let's say lower priced subscription plan is showing the most fluctuation and there are only two types of plans
In this subscription plan which age group is showing the most fluctuation (18-24,25-30, 30+ etc) ?
Is there any seasonality trend observed (eg: placemen"See full answer
Data Scientist
Technical
🧠 Want an expert answer to a question? Saving questions lets us know what content to make next.
"from typing import List
def traprainwater(height: List[int]) -> int:
if not height:
return 0
l, r = 0, len(height) - 1
leftMax, rightMax = height[l], height[r]
res = 0
while l < r:
if leftMax < rightMax:
l += 1
leftMax = max(leftMax, height[l])
res += leftMax - height[l]
else:
r -= 1
rightMax = max(rightMax, height[r])
"
Anonymous Roadrunner - "from typing import List
def traprainwater(height: List[int]) -> int:
if not height:
return 0
l, r = 0, len(height) - 1
leftMax, rightMax = height[l], height[r]
res = 0
while l < r:
if leftMax < rightMax:
l += 1
leftMax = max(leftMax, height[l])
res += leftMax - height[l]
else:
r -= 1
rightMax = max(rightMax, height[r])
"See full answer
"To handle the non-uniform sampling, I'd first clean and divide the dataset into chunks of n second interval 'uniform' trajectory data(e.g. 5s or 10s trajectories). This gives us a cleaner trajectory data chunks, T, of format (ship_ID, x, y, z, timestamp) to be formed.
For the system itself, I'd use a generative model, e.g. Variational AutoEncoder (VAE), and train the model's 'encoder' to produce a latent-space representation of input features (x,y,z,timestamp) from T, and it's 'decoder' to pred"
Anonymous Hornet - "To handle the non-uniform sampling, I'd first clean and divide the dataset into chunks of n second interval 'uniform' trajectory data(e.g. 5s or 10s trajectories). This gives us a cleaner trajectory data chunks, T, of format (ship_ID, x, y, z, timestamp) to be formed.
For the system itself, I'd use a generative model, e.g. Variational AutoEncoder (VAE), and train the model's 'encoder' to produce a latent-space representation of input features (x,y,z,timestamp) from T, and it's 'decoder' to pred"See full answer
"I don't have experience working with alot of Biological Scientists. Most of my experience comes with Data Scientists. Described how I used ideation techniques like brainstorming and other creative ways to get people to find common ground. I also mentioned how I like to do survey's before meetings to prompt people and also get unbiased opnions"
Mark M. - "I don't have experience working with alot of Biological Scientists. Most of my experience comes with Data Scientists. Described how I used ideation techniques like brainstorming and other creative ways to get people to find common ground. I also mentioned how I like to do survey's before meetings to prompt people and also get unbiased opnions"See full answer
"The cases where data is under heavy outlier influence. Since mean fluctuates due to the presence of an outlier, median might be a better measure"
Himani E. - "The cases where data is under heavy outlier influence. Since mean fluctuates due to the presence of an outlier, median might be a better measure"See full answer
"I’ve spent over 6 years building and scaling e-commerce products across EMEA and APAC.
At Jumia, I led product initiatives on the checkout and payments side. For example, I launched gamified promotions on PDP and checkout that improved engagement and delivered a 2.3x uplift in conversion. I also introduced automated installment payments and order cancellation flows, which not only improved user trust but also reduced complaints by 30% and lowered operational costs.
Before that, at Lazada, I work"
Rajeev K. - "I’ve spent over 6 years building and scaling e-commerce products across EMEA and APAC.
At Jumia, I led product initiatives on the checkout and payments side. For example, I launched gamified promotions on PDP and checkout that improved engagement and delivered a 2.3x uplift in conversion. I also introduced automated installment payments and order cancellation flows, which not only improved user trust but also reduced complaints by 30% and lowered operational costs.
Before that, at Lazada, I work"See full answer
"A Random Forest works by building an ensemble of decision trees, each trained on a slightly different version of the data. The key mechanism is bagging: for each tree, we sample the training data with replacement (bootstrapping), so every tree sees a different subset of examples. On top of that, at each split the algorithm randomly selects a subset of features, so trees explore different predictors.
These two sources of randomness decorrelate the trees. When we aggregate them — by averag"
Yuexiang Y. - "A Random Forest works by building an ensemble of decision trees, each trained on a slightly different version of the data. The key mechanism is bagging: for each tree, we sample the training data with replacement (bootstrapping), so every tree sees a different subset of examples. On top of that, at each split the algorithm randomly selects a subset of features, so trees explore different predictors.
These two sources of randomness decorrelate the trees. When we aggregate them — by averag"See full answer
"[I'm not sure whether the answer below is the best, as I have not gotten result and feedback from my interview]
Ans: I would solve by first using a VAE-style model, to create a latent space embedding that translates user description to generate images. Training would be done on the 1000 avatar images and 100000 descriptions, following this scheme:
VAE:
description -> encoder -> latent space -> decoder -> image
Q: "OK, but that means you're limiting the generated images to be only the 1000 imag"
Nick S. - "[I'm not sure whether the answer below is the best, as I have not gotten result and feedback from my interview]
Ans: I would solve by first using a VAE-style model, to create a latent space embedding that translates user description to generate images. Training would be done on the 1000 avatar images and 100000 descriptions, following this scheme:
VAE:
description -> encoder -> latent space -> decoder -> image
Q: "OK, but that means you're limiting the generated images to be only the 1000 imag"See full answer