abordado_ has joined #lima
abordado_ has quit [Remote host closed the connection]
yuq825 has joined #lima
yuq825 has quit [Ping timeout: 258 seconds]
yuq825 has joined #lima
adjtm has quit [Ping timeout: 240 seconds]
<
yuq825>
from the cpu side, supertuxkart spend 22.67% time on u_vbuf_get_minmax_index_mapped
<
yuq825>
the index buffer is write combined, so read is expensive, not mention calculate minmax every time
<
yuq825>
mem copy rate 2.319644 GB/s
<
yuq825>
gbm copy rate 0.048172 GB/s
<
yuq825>
read speed of gpu buffer is slow
megi has quit [Ping timeout: 272 seconds]
<
anarsoul|c>
yuq825: Alyssa implemented minmax caching for panfrost
<
anarsoul|c>
I guess we can do the same
buzzmarshall has quit [Remote host closed the connection]
Barada has joined #lima
enunes has quit [Ping timeout: 260 seconds]
enunes has joined #lima
<
anarsoul>
I'm trying to revive Connor's gpir optimization MR
<
anarsoul>
lima_nir_split_loads.c is broken if ssa has several successors
<
anarsoul>
well, several successors from different blocks
<
rellla>
anarsoul: thanks for looking into gpir. i'm thinking about your comment on my MR, but i didn't get it yet...
<
anarsoul>
dEQP-GLES2.functional.shaders.indexing.matrix_subscript.mat2_dynamic_loop_write_dynamic_read_vertex results in infinite loop :(
<
anarsoul>
in gpir compiler
dddddd has quit [Ping timeout: 272 seconds]
<
anarsoul>
it stucks in scheduler
<
anarsoul>
rellla: btw I barely understand gpir code :)
<
anarsoul>
it's pretty complex
<
anarsoul>
it ends up with 2 ld_reg that it can't schedule :\
<
anarsoul>
rellla: we should also align it to 16*16
<
anarsoul>
otherwise 2nd cubemap face may have unaligned address
<
rellla>
anarsoul: sorry, but i still don't get it :p what should be aligned? (size * pres->array_size)?
<
anarsoul>
look above
<
anarsoul>
find "size += actual_level_size; /* Save some memory */"
<
anarsoul>
note that for last mipmap level it doesn't do alignment
<
anarsoul>
last mipmap level is not last anymore if array_size > 1
<
rellla>
yes, understand that, but why 256 and not 64?
<
anarsoul>
because you have to align dimensions to 16
<
anarsoul>
each dimension
<
anarsoul>
technically it will break even if array_size is 1 and you're trying to render into the last miplevel :)
adjtm has joined #lima
<
rellla>
why isn't this a problem for the !last_levels?
<
anarsoul>
look above
<
anarsoul>
in the code
<
anarsoul>
oh, layer_stride is already aligned
<
anarsoul>
so yeah, here you need to align it to 64
<
rellla>
so it should be "size += align(actual_level_size, 64); /* Don't save some memory */"
<
anarsoul>
it's actually broken for formats different from rgba8888
<
anarsoul>
it shouldn't be size += 0x0400;
<
anarsoul>
it should be `else size += util_format_get_stride(pres->format, 16, 16);`
<
rellla>
util_format_get_stride only takes 2 arguments :)
gcl_ has quit [Ping timeout: 240 seconds]
gcl has joined #lima
kaspter has joined #lima
gcl_ has joined #lima
gcl has quit [Ping timeout: 255 seconds]
yann has quit [Ping timeout: 258 seconds]
<
rellla>
anarsoul: seems that dEQP-GLES2.functional.default_vertex_attrib.* are all mostly flakes :(
<
rellla>
anarsoul: can i keep your r-b?
yann has joined #lima
_whitelogger has joined #lima
megi has joined #lima
kaspter has quit [Ping timeout: 260 seconds]
yuq825 has quit [Ping timeout: 265 seconds]
yuq825 has joined #lima
Barada has quit [Quit: Barada]
Barada has joined #lima
Barada has quit [Remote host closed the connection]
Barada has joined #lima
yuq825 has quit [Quit: Leaving.]
Barada has quit [Quit: Barada]
Barada has joined #lima
dddddd has joined #lima
_whitelogger has joined #lima
Barada has quit [Quit: Barada]
drod has joined #lima
drod has quit [Ping timeout: 272 seconds]
drod has joined #lima
yann has quit [Ping timeout: 260 seconds]
drod has quit [Read error: Connection reset by peer]
gcl_ has quit [Ping timeout: 255 seconds]
gcl has joined #lima
yann has joined #lima
buzzmarshall has joined #lima
gcl_ has joined #lima
gcl has quit [Ping timeout: 240 seconds]
_whitelogger has joined #lima