parabellum630

parabellum630 OP t1_ix5yyv8 wrote

Oh my God. I used to do this too! I am happy I am not the only one!! But my monkey brain learned not to do this eventually. I have managed to get it to GRU performance by applying more warmup steps, learning rate scheduling, decreasing model size, using Pre-LN, doubling the batch size, and reducing the sequence length.

2

parabellum630 OP t1_ix4fc49 wrote

Thank you!! I was experimenting with off-the-shelf implementation with little customization. I am using the transformer in an encoder fashion with 800 hidden dimensions due to the constraints of other models surrounding it. I will try out varying all these hyper parameters. Looks like it's going to be a long week.

2

parabellum630 t1_itmm0ei wrote

I worked on eyes and upper head shape while the paper I built upon (DECA) tackles personalized fine details. But I do agree hair modeling and finer details and texturing is better with volumetric approaches. Maybe a fusion of the two would be a good research direction? Like the end goal is not only the modeling but also using it for real tasks.

2

parabellum630 t1_itmihvd wrote

What are your thoughts about 3DMM based models. I am doing research on avatar generation from monocular images using 3DMM based approaches and found them to be better in terms of inference speed, animatability and ease of integration into existing creative workflows.

2