diff --git "a/MI300x_llama8b_pp1_tp1/profiling_bs1_pl256.json" "b/MI300x_llama8b_pp1_tp1/profiling_bs1_pl256.json" --- "a/MI300x_llama8b_pp1_tp1/profiling_bs1_pl256.json" +++ "b/MI300x_llama8b_pp1_tp1/profiling_bs1_pl256.json" @@ -130,24 +130,24 @@ { "entry": { "name": "LlamaForCausalLM", - "cuda_time_us": 18191.011, - "pct_cuda_time": 97.49206882589682, + "cuda_time_us": 18189.998000000007, + "pct_cuda_time": 97.51162527398678, "invocations": 1 }, "children": [ { "entry": { "name": "VocabParallelEmbedding(weight=bfloat16[128256, 4096])", - "cuda_time_us": 7.21, - "pct_cuda_time": 0.03864094283900527, + "cuda_time_us": 7.85, + "pct_cuda_time": 0.04208171207060033, "invocations": 1 }, "children": [ { "entry": { "name": "void at::native::(anonymous namespace)::indexSelectLargeIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, unsigned int, long)", - "cuda_time_us": 7.21, - "pct_cuda_time": 0.03864094283900527, + "cuda_time_us": 7.85, + "pct_cuda_time": 0.04208171207060033, "invocations": 1 }, "children": [] @@ -157,24 +157,24 @@ { "entry": { "name": "LlamaDecoderLayer", - "cuda_time_us": 18173.827, - "pct_cuda_time": 97.39997368557152, + "cuda_time_us": 18172.214000000004, + "pct_cuda_time": 97.41629009341816, "invocations": 32 }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cuda_time_us": 622.5070000000002, - "pct_cuda_time": 3.3362354235618112, + "cuda_time_us": 625.6349999999999, + "pct_cuda_time": 3.353858844750323, "invocations": 64 }, "children": [ { "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::rms_norm_kernel(c10::BFloat16*, c10::BFloat16 const*, c10::BFloat16 const*, float, int, unsigned long, unsigned long)", - "cuda_time_us": 5.166, - "pct_cuda_time": 0.027686423121539704, + "cuda_time_us": 5.607, + "pct_cuda_time": 0.030057599946478483, "invocations": 1 }, "children": [] @@ -182,8 +182,8 @@ { "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", - "cuda_time_us": 617.3410000000001, - "pct_cuda_time": 3.3085490004402716, + "cuda_time_us": 620.0279999999998, + "pct_cuda_time": 3.3238012448038443, "invocations": 63 }, "children": [] @@ -193,24 +193,24 @@ { "entry": { "name": "LlamaAttention", - "cuda_time_us": 6983.4929999999995, - "pct_cuda_time": 37.42701162685068, + "cuda_time_us": 7004.768000000001, + "pct_cuda_time": 37.550653515586625, "invocations": 32 }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cuda_time_us": 1738.7930000000006, - "pct_cuda_time": 9.318807340064149, + "cuda_time_us": 1731.764, + "pct_cuda_time": 9.283515162067658, "invocations": 32 }, "children": [ { "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", - "cuda_time_us": 1738.7930000000006, - "pct_cuda_time": 9.318807340064149, + "cuda_time_us": 1731.764, + "pct_cuda_time": 9.283515162067658, "invocations": 32 }, "children": [] @@ -220,16 +220,16 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cuda_time_us": 857.7579999999999, - "pct_cuda_time": 4.597028827697571, + "cuda_time_us": 844.9809999999999, + "pct_cuda_time": 4.529713012372985, "invocations": 32 }, "children": [ { "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 629.3709999999998, - "pct_cuda_time": 3.373022029892868, + "cuda_time_us": 617.0329999999999, + "pct_cuda_time": 3.307745865485189, "invocations": 128 }, "children": [] @@ -237,8 +237,8 @@ { "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", - "cuda_time_us": 228.38700000000006, - "pct_cuda_time": 1.2240067978047018, + "cuda_time_us": 227.94800000000006, + "pct_cuda_time": 1.2219671468877973, "invocations": 32 }, "children": [] @@ -248,16 +248,16 @@ { "entry": { "name": "Attention", - "cuda_time_us": 3009.6349999999998, - "pct_cuda_time": 16.129699584087327, + "cuda_time_us": 3051.023, + "pct_cuda_time": 16.355703363920924, "invocations": 32 }, "children": [ { "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", - "cuda_time_us": 179.31999999999996, - "pct_cuda_time": 0.9610393716907661, + "cuda_time_us": 188.332, + "pct_cuda_time": 1.0095965602140513, "invocations": 32 }, "children": [] @@ -265,8 +265,8 @@ { "entry": { "name": "attn_fwd", - "cuda_time_us": 2830.3149999999996, - "pct_cuda_time": 15.16866021239656, + "cuda_time_us": 2862.6910000000003, + "pct_cuda_time": 15.346106803706874, "invocations": 32 }, "children": [] @@ -276,16 +276,16 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cuda_time_us": 1377.307, - "pct_cuda_time": 7.381475875001642, + "cuda_time_us": 1377.0, + "pct_cuda_time": 7.3817219772250535, "invocations": 32 }, "children": [ { "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", - "cuda_time_us": 1377.307, - "pct_cuda_time": 7.381475875001642, + "cuda_time_us": 1377.0, + "pct_cuda_time": 7.3817219772250535, "invocations": 32 }, "children": [] @@ -297,24 +297,24 @@ { "entry": { "name": "LlamaMLP", - "cuda_time_us": 10567.827000000001, - "pct_cuda_time": 56.63672663515903, + "cuda_time_us": 10541.811, + "pct_cuda_time": 56.51177773308119, "invocations": 32 }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cuda_time_us": 6272.113, - "pct_cuda_time": 33.61447432909596, + "cuda_time_us": 6245.850000000001, + "pct_cuda_time": 33.48230080715403, "invocations": 32 }, "children": [ { "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", - "cuda_time_us": 6272.113, - "pct_cuda_time": 33.61447432909596, + "cuda_time_us": 6245.850000000001, + "pct_cuda_time": 33.48230080715403, "invocations": 32 }, "children": [] @@ -324,16 +324,16 @@ { "entry": { "name": "SiluAndMul", - "cuda_time_us": 456.39799999999997, - "pct_cuda_time": 2.4459984784793796, + "cuda_time_us": 462.576, + "pct_cuda_time": 2.4797439544929962, "invocations": 32 }, "children": [ { "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", - "cuda_time_us": 456.39799999999997, - "pct_cuda_time": 2.4459984784793796, + "cuda_time_us": 462.576, + "pct_cuda_time": 2.4797439544929962, "invocations": 32 }, "children": [] @@ -343,16 +343,16 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cuda_time_us": 3839.316000000001, - "pct_cuda_time": 20.57625382758369, + "cuda_time_us": 3833.3849999999998, + "pct_cuda_time": 20.549732971434175, "invocations": 32 }, "children": [ { "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", - "cuda_time_us": 3630.76, - "pct_cuda_time": 19.45852838032549, + "cuda_time_us": 3625.067, + "pct_cuda_time": 19.43299690836114, "invocations": 32 }, "children": [] @@ -360,8 +360,8 @@ { "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", - "cuda_time_us": 208.556, - "pct_cuda_time": 1.1177254472581948, + "cuda_time_us": 208.31800000000004, + "pct_cuda_time": 1.1167360630730347, "invocations": 32 }, "children": [] @@ -375,16 +375,16 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cuda_time_us": 9.974, - "pct_cuda_time": 0.05345419748630216, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "invocations": 1 }, "children": [ { "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", - "cuda_time_us": 9.974, - "pct_cuda_time": 0.05345419748630216, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "invocations": 1 }, "children": [] @@ -396,16 +396,16 @@ { "entry": { "name": "LogitsProcessor", - "cuda_time_us": 241.549, - "pct_cuda_time": 1.2945466160636456, + "cuda_time_us": 241.62900000000002, + "pct_cuda_time": 1.2953072618989923, "invocations": 1 }, "children": [ { "entry": { "name": "void at::native::(anonymous namespace)::indexSelectSmallIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, long)", - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.085, + "pct_cuda_time": 0.021898572459669093, "invocations": 1 }, "children": [] @@ -413,8 +413,8 @@ { "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", - "cuda_time_us": 236.983, - "pct_cuda_time": 1.270075805383632, + "cuda_time_us": 237.544, + "pct_cuda_time": 1.2734086894393233, "invocations": 1 }, "children": [] @@ -424,16 +424,16 @@ { "entry": { "name": "Sampler", - "cuda_time_us": 226.40500000000003, - "pct_cuda_time": 1.213384558039527, + "cuda_time_us": 222.557, + "pct_cuda_time": 1.1930674641142165, "invocations": 1 }, "children": [ { "entry": { "name": "Memcpy DtoD (Device -> Device)", - "cuda_time_us": 33.443, - "pct_cuda_time": 0.17923287813659544, + "cuda_time_us": 29.957, + "pct_cuda_time": 0.16059131828012413, "invocations": 7 }, "children": [] @@ -441,8 +441,8 @@ { "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 2.443, - "pct_cuda_time": 0.013092901991080427, + "cuda_time_us": 3.524, + "pct_cuda_time": 0.01889120424672555, "invocations": 1 }, "children": [] @@ -450,8 +450,8 @@ { "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2}>(int, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2})", - "cuda_time_us": 14.141, - "pct_cuda_time": 0.07578662589270091, + "cuda_time_us": 13.94, + "pct_cuda_time": 0.0747285434731425, "invocations": 1 }, "children": [] @@ -459,8 +459,8 @@ { "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::SoftMaxForwardEpilogue>(float*, float const*, int)", - "cuda_time_us": 70.985, - "pct_cuda_time": 0.3804337486028834, + "cuda_time_us": 71.145, + "pct_cuda_time": 0.38138896882329437, "invocations": 1 }, "children": [] @@ -468,8 +468,8 @@ { "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::LogSoftMaxForwardEpilogue>(float*, float const*, int)", - "cuda_time_us": 40.42, - "pct_cuda_time": 0.2166250914774748, + "cuda_time_us": 40.259, + "pct_cuda_time": 0.21581753455417824, "invocations": 1 }, "children": [] @@ -477,8 +477,8 @@ { "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "invocations": 1 }, "children": [] @@ -486,8 +486,8 @@ { "entry": { "name": "void at::native::index_elementwise_kernel<128, 4, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1}>(long, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 8.251, - "pct_cuda_time": 0.04422003042505305, + "cuda_time_us": 6.73, + "pct_cuda_time": 0.03607769710001787, "invocations": 1 }, "children": [] @@ -495,8 +495,8 @@ { "entry": { "name": "void at::native::reduce_kernel<512, 1, at::native::ReduceOp, unsigned int, long, 4, 4> >(at::native::ReduceOp, unsigned int, long, 4, 4>)", - "cuda_time_us": 46.949, - "pct_cuda_time": 0.2516163141953479, + "cuda_time_us": 47.63, + "pct_cuda_time": 0.2553314580793241, "invocations": 1 }, "children": [] @@ -504,8 +504,8 @@ { "entry": { "name": "Memcpy DtoH (Device -> Host)", - "cuda_time_us": 5.007, - "pct_cuda_time": 0.02683428582453528, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "invocations": 1 }, "children": [] @@ -517,18 +517,18 @@ { "entry": { "name": "LlamaForCausalLM", - "cpu_time_us": 34566.393, - "cuda_time_us": 18191.011, - "pct_cuda_time": 97.49206882589682, + "cpu_time_us": 35864.356, + "cuda_time_us": 18189.998000000007, + "pct_cuda_time": 97.51162527398678, "trace": "" }, "children": [ { "entry": { "name": "VocabParallelEmbedding(weight=bfloat16[128256, 4096])", - "cpu_time_us": 66.9, - "cuda_time_us": 7.21, - "pct_cuda_time": 0.03864094283900527, + "cpu_time_us": 54.602, + "cuda_time_us": 7.85, + "pct_cuda_time": 0.04208171207060033, "trace": "" }, "children": [ @@ -536,8 +536,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::indexSelectLargeIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, unsigned int, long)", "cpu_time_us": 0, - "cuda_time_us": 7.21, - "pct_cuda_time": 0.03864094283900527, + "cuda_time_us": 7.85, + "pct_cuda_time": 0.04208171207060033, "trace": "index_select(bfloat16[128256, 4096], 0, int64[256]) <- embedding(bfloat16[128256, 4096], int64[256], -1, False, False)" }, "children": [] @@ -547,18 +547,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1477.095, - "cuda_time_us": 577.6560000000001, - "pct_cuda_time": 3.095863034203666, + "cpu_time_us": 1688.925, + "cuda_time_us": 586.312, + "pct_cuda_time": 3.1430589512787037, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 78.106, - "cuda_time_us": 5.166, - "pct_cuda_time": 0.027686423121539704, + "cpu_time_us": 73.41, + "cuda_time_us": 5.607, + "pct_cuda_time": 0.030057599946478483, "trace": "" }, "children": [ @@ -566,8 +566,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::rms_norm_kernel(c10::BFloat16*, c10::BFloat16 const*, c10::BFloat16 const*, float, int, unsigned long, unsigned long)", "cpu_time_us": 0, - "cuda_time_us": 5.166, - "pct_cuda_time": 0.027686423121539704, + "cuda_time_us": 5.607, + "pct_cuda_time": 0.030057599946478483, "trace": "_C::rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -577,18 +577,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 1073.462, - "cuda_time_us": 219.038, - "pct_cuda_time": 1.1739021966116556, + "cpu_time_us": 1267.324, + "cuda_time_us": 231.307, + "pct_cuda_time": 1.2399738310718917, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 175.832, - "cuda_time_us": 60.218, - "pct_cuda_time": 0.3227295833396976, + "cpu_time_us": 174.521, + "cuda_time_us": 58.246, + "pct_cuda_time": 0.3122409428362022, "trace": "" }, "children": [ @@ -596,8 +596,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 60.218, - "pct_cuda_time": 0.3227295833396976, + "cuda_time_us": 58.246, + "pct_cuda_time": 0.3122409428362022, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -607,9 +607,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 95.663, - "cuda_time_us": 25.232999999999997, - "pct_cuda_time": 0.13523258122838, + "cpu_time_us": 103.124, + "cuda_time_us": 27.716000000000005, + "pct_cuda_time": 0.1485779276113069, "trace": "" }, "children": [ @@ -617,8 +617,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.046, - "pct_cuda_time": 0.027043300633234478, + "cuda_time_us": 5.447, + "pct_cuda_time": 0.02919988352210956, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -627,8 +627,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.046, - "pct_cuda_time": 0.02168394656402432, + "cuda_time_us": 4.605, + "pct_cuda_time": 0.024686150838868093, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -637,8 +637,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.33, - "pct_cuda_time": 0.03928406532731049, + "cuda_time_us": 8.051, + "pct_cuda_time": 0.0431592183287138, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -647,8 +647,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.487, - "pct_cuda_time": 0.029406775777756165, + "cuda_time_us": 6.048, + "pct_cuda_time": 0.03242168084114533, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -657,8 +657,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 3.324, - "pct_cuda_time": 0.01781449292605458, + "cuda_time_us": 3.565, + "pct_cuda_time": 0.019110994080470088, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -668,9 +668,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 632.474, - "cuda_time_us": 90.614, - "pct_cuda_time": 0.4856325096274097, + "cpu_time_us": 802.45, + "cuda_time_us": 90.774, + "pct_cuda_time": 0.48661469191040446, "trace": "" }, "children": [ @@ -678,8 +678,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.808, - "pct_cuda_time": 0.031127128433972626, + "cuda_time_us": 5.888, + "pct_cuda_time": 0.031563964416776405, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -688,8 +688,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 84.806, - "pct_cuda_time": 0.454505381193437, + "cuda_time_us": 84.886, + "pct_cuda_time": 0.455050727493628, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -699,9 +699,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 115.382, - "cuda_time_us": 42.973, - "pct_cuda_time": 0.2303075224161683, + "cpu_time_us": 129.574, + "cuda_time_us": 54.571, + "pct_cuda_time": 0.2925402687139784, "trace": "" }, "children": [ @@ -709,8 +709,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.973, - "pct_cuda_time": 0.2303075224161683, + "cuda_time_us": 54.571, + "pct_cuda_time": 0.2925402687139784, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -722,9 +722,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 47.071, - "cuda_time_us": 9.693, - "pct_cuda_time": 0.0519482189928541, + "cpu_time_us": 46.429, + "cuda_time_us": 10.014, + "pct_cuda_time": 0.05368232671019003, "trace": "" }, "children": [ @@ -732,8 +732,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.693, - "pct_cuda_time": 0.0519482189928541, + "cuda_time_us": 10.014, + "pct_cuda_time": 0.05368232671019003, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -743,18 +743,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 248.731, - "cuda_time_us": 343.759, - "pct_cuda_time": 1.8423261954776164, + "cpu_time_us": 272.087, + "cuda_time_us": 339.384, + "pct_cuda_time": 1.8193451935501432, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 94.331, - "cuda_time_us": 201.63, - "pct_cuda_time": 1.080606560974845, + "cpu_time_us": 100.971, + "cuda_time_us": 201.2, + "pct_cuda_time": 1.0785784036439219, "trace": "" }, "children": [ @@ -762,8 +762,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 201.63, - "pct_cuda_time": 1.080606560974845, + "cuda_time_us": 201.2, + "pct_cuda_time": 1.0785784036439219, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -773,9 +773,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 32.409, - "cuda_time_us": 15.132, - "pct_cuda_time": 0.08109774577528818, + "cpu_time_us": 33.911, + "cuda_time_us": 15.222, + "pct_cuda_time": 0.08160099632339851, "trace": "" }, "children": [ @@ -783,8 +783,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.132, - "pct_cuda_time": 0.08109774577528818, + "cuda_time_us": 15.222, + "pct_cuda_time": 0.08160099632339851, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -794,9 +794,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 99.368, - "cuda_time_us": 126.997, - "pct_cuda_time": 0.680621888727483, + "cpu_time_us": 106.51, + "cuda_time_us": 122.96199999999999, + "pct_cuda_time": 0.6591657935828227, "trace": "" }, "children": [ @@ -804,8 +804,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 120.108, - "pct_cuda_time": 0.6437012985446943, + "cuda_time_us": 116.353, + "pct_cuda_time": 0.6237367445287338, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -814,8 +814,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.889, - "pct_cuda_time": 0.03692059018278881, + "cuda_time_us": 6.609, + "pct_cuda_time": 0.03542904905408887, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -829,18 +829,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1081.305, - "cuda_time_us": 573.1500000000001, - "pct_cuda_time": 3.0717137847678053, + "cpu_time_us": 1096.67, + "cuda_time_us": 586.433, + "pct_cuda_time": 3.1437075993246326, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.027, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cpu_time_us": 36.314, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "" }, "children": [ @@ -848,8 +848,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -859,18 +859,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 757.28, - "cuda_time_us": 214.901, - "pct_cuda_time": 1.1517305488273333, + "cpu_time_us": 765.574, + "cuda_time_us": 228.584, + "pct_cuda_time": 1.2253765696746635, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 90.395, - "cuda_time_us": 57.095, - "pct_cuda_time": 0.30599232058155423, + "cpu_time_us": 85.778, + "cuda_time_us": 56.644, + "pct_cuda_time": 0.3036530571372083, "trace": "" }, "children": [ @@ -878,8 +878,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 57.095, - "pct_cuda_time": 0.30599232058155423, + "cuda_time_us": 56.644, + "pct_cuda_time": 0.3036530571372083, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -889,9 +889,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 79.95, - "cuda_time_us": 25.032000000000004, - "pct_cuda_time": 0.1341553510604688, + "cpu_time_us": 79.329, + "cuda_time_us": 24.712, + "pct_cuda_time": 0.1324743017437803, "trace": "" }, "children": [ @@ -899,8 +899,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.485, - "pct_cuda_time": 0.02403670300040758, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -910,7 +910,7 @@ "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, "cuda_time_us": 4.406, - "pct_cuda_time": 0.023613314028939975, + "pct_cuda_time": 0.02361936603605924, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -919,8 +919,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.288, - "pct_cuda_time": 0.0336996183871935, + "cuda_time_us": 6.328, + "pct_cuda_time": 0.03392268458379095, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -929,8 +929,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.808, - "pct_cuda_time": 0.031127128433972626, + "cuda_time_us": 5.407, + "pct_cuda_time": 0.02898545441601733, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -939,8 +939,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.021678587209955108, + "cuda_time_us": 4.046, + "pct_cuda_time": 0.021689504081229168, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -950,9 +950,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 446.306, - "cuda_time_us": 90.733, - "pct_cuda_time": 0.4862702727616457, + "cpu_time_us": 457.283, + "cuda_time_us": 91.255, + "pct_cuda_time": 0.4891932019111635, "trace": "" }, "children": [ @@ -960,8 +960,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.168, - "pct_cuda_time": 0.03305649589888828, + "cuda_time_us": 6.529, + "pct_cuda_time": 0.03500019084190441, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -970,8 +970,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 84.565, - "pct_cuda_time": 0.4532137768627573, + "cuda_time_us": 84.726, + "pct_cuda_time": 0.4541930110692591, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -981,9 +981,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 102.924, - "cuda_time_us": 42.041, - "pct_cuda_time": 0.2253126044236644, + "cpu_time_us": 104.566, + "cuda_time_us": 55.973, + "pct_cuda_time": 0.30005600888251116, "trace": "" }, "children": [ @@ -991,8 +991,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.041, - "pct_cuda_time": 0.2253126044236644, + "cuda_time_us": 55.973, + "pct_cuda_time": 0.30005600888251116, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -1004,9 +1004,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.947, - "cuda_time_us": 9.453, - "pct_cuda_time": 0.05066197401624366, + "cpu_time_us": 40.44, + "cuda_time_us": 10.134, + "pct_cuda_time": 0.05432561402846672, "trace": "" }, "children": [ @@ -1014,8 +1014,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.453, - "pct_cuda_time": 0.05066197401624366, + "cuda_time_us": 10.134, + "pct_cuda_time": 0.05432561402846672, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1025,18 +1025,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 223.474, - "cuda_time_us": 339.463, - "pct_cuda_time": 1.8193024103962896, + "cpu_time_us": 227.98, + "cuda_time_us": 338.302, + "pct_cuda_time": 1.8135448862303485, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 87.891, - "cuda_time_us": 199.267, - "pct_cuda_time": 1.0679424073093013, + "cpu_time_us": 87.951, + "cuda_time_us": 199.207, + "pct_cuda_time": 1.0678944734328766, "trace": "" }, "children": [ @@ -1044,8 +1044,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 199.267, - "pct_cuda_time": 1.0679424073093013, + "cuda_time_us": 199.207, + "pct_cuda_time": 1.0678944734328766, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -1055,9 +1055,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.03, - "cuda_time_us": 14.841, - "pct_cuda_time": 0.07953817374114801, + "cpu_time_us": 28.262, + "cuda_time_us": 16.584, + "pct_cuda_time": 0.08890230738583897, "trace": "" }, "children": [ @@ -1065,8 +1065,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 14.841, - "pct_cuda_time": 0.07953817374114801, + "cuda_time_us": 16.584, + "pct_cuda_time": 0.08890230738583897, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -1076,9 +1076,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 91.357, - "cuda_time_us": 125.355, - "pct_cuda_time": 0.67182182934584, + "cpu_time_us": 92.899, + "cuda_time_us": 122.511, + "pct_cuda_time": 0.6567481054116328, "trace": "" }, "children": [ @@ -1086,8 +1086,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 118.506, - "pct_cuda_time": 0.6351156133258196, + "cuda_time_us": 116.222, + "pct_cuda_time": 0.6230344892062817, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1096,8 +1096,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.849, - "pct_cuda_time": 0.0367062160200204, + "cuda_time_us": 6.289, + "pct_cuda_time": 0.033713616205351016, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1111,18 +1111,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1058.56, - "cuda_time_us": 585.37, - "pct_cuda_time": 3.137205091493553, + "cpu_time_us": 1070.34, + "cuda_time_us": 576.9279999999999, + "pct_cuda_time": 3.0927538829894656, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.17, - "cuda_time_us": 9.373, - "pct_cuda_time": 0.05023322569070685, + "cpu_time_us": 35.343, + "cuda_time_us": 9.253, + "pct_cuda_time": 0.04960281296678534, "trace": "" }, "children": [ @@ -1130,8 +1130,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.373, - "pct_cuda_time": 0.05023322569070685, + "cuda_time_us": 9.253, + "pct_cuda_time": 0.04960281296678534, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1141,18 +1141,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 727.636, - "cuda_time_us": 217.20600000000002, - "pct_cuda_time": 1.1640838599568626, + "cpu_time_us": 745.654, + "cuda_time_us": 213.88099999999997, + "pct_cuda_time": 1.1465577910028113, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.964, - "cuda_time_us": 55.031, - "pct_cuda_time": 0.29493061378270446, + "cpu_time_us": 89.644, + "cuda_time_us": 56.423, + "pct_cuda_time": 0.30246833632604875, "trace": "" }, "children": [ @@ -1160,8 +1160,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 55.031, - "pct_cuda_time": 0.29493061378270446, + "cuda_time_us": 56.423, + "pct_cuda_time": 0.30246833632604875, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -1171,9 +1171,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.52, - "cuda_time_us": 26.154, - "pct_cuda_time": 0.1401685463261226, + "cpu_time_us": 77.777, + "cuda_time_us": 26.314, + "pct_cuda_time": 0.14106218744277418, "trace": "" }, "children": [ @@ -1182,7 +1182,7 @@ "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "pct_cuda_time": 0.0246915115665204, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -1191,8 +1191,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.446, - "pct_cuda_time": 0.02382768819170838, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -1201,8 +1201,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.489, - "pct_cuda_time": 0.034776848555104746, + "cuda_time_us": 6.649, + "pct_cuda_time": 0.0356434781601811, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -1211,8 +1211,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.767, - "pct_cuda_time": 0.030907394917135008, + "cuda_time_us": 5.887, + "pct_cuda_time": 0.031558603689124096, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -1221,8 +1221,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.02597142981939245, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.02490594067261263, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -1232,9 +1232,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 435.26, - "cuda_time_us": 90.614, - "pct_cuda_time": 0.4856325096274097, + "cpu_time_us": 440.539, + "cuda_time_us": 91.29499999999999, + "pct_cuda_time": 0.48940763101725565, "trace": "" }, "children": [ @@ -1242,8 +1242,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.528, - "pct_cuda_time": 0.029626509294593775, + "cuda_time_us": 5.448, + "pct_cuda_time": 0.02920524424976187, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -1252,8 +1252,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 85.086, - "pct_cuda_time": 0.45600600033281585, + "cuda_time_us": 85.847, + "pct_cuda_time": 0.4602023867674938, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -1263,9 +1263,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 99.899, - "cuda_time_us": 45.407, - "pct_cuda_time": 0.24335219022062582, + "cpu_time_us": 101.092, + "cuda_time_us": 39.849, + "pct_cuda_time": 0.21361963621673283, "trace": "" }, "children": [ @@ -1273,8 +1273,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 45.407, - "pct_cuda_time": 0.24335219022062582, + "cuda_time_us": 39.849, + "pct_cuda_time": 0.21361963621673283, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -1286,9 +1286,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 40.24, - "cuda_time_us": 9.894, - "pct_cuda_time": 0.053025449160765355, + "cpu_time_us": 38.267, + "cuda_time_us": 9.813, + "pct_cuda_time": 0.052604820452076574, "trace": "" }, "children": [ @@ -1296,8 +1296,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.894, - "pct_cuda_time": 0.053025449160765355, + "cuda_time_us": 9.813, + "pct_cuda_time": 0.052604820452076574, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1307,18 +1307,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 229.863, - "cuda_time_us": 348.897, - "pct_cuda_time": 1.869862556685218, + "cpu_time_us": 225.877, + "cuda_time_us": 343.981, + "pct_cuda_time": 1.843988458567793, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 90.786, - "cuda_time_us": 208.911, - "pct_cuda_time": 1.1196280179527642, + "cpu_time_us": 89.403, + "cuda_time_us": 205.627, + "pct_cuda_time": 1.1023103449606797, "trace": "" }, "children": [ @@ -1326,8 +1326,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 208.911, - "pct_cuda_time": 1.1196280179527642, + "cuda_time_us": 205.627, + "pct_cuda_time": 1.1023103449606797, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -1337,9 +1337,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.19, - "cuda_time_us": 15.422, - "pct_cuda_time": 0.08265195845535912, + "cpu_time_us": 27.23, + "cuda_time_us": 14.941, + "pct_cuda_time": 0.08009463185310059, "trace": "" }, "children": [ @@ -1347,8 +1347,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.422, - "pct_cuda_time": 0.08265195845535912, + "cuda_time_us": 14.941, + "pct_cuda_time": 0.08009463185310059, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -1358,9 +1358,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 93.37, - "cuda_time_us": 124.56400000000001, - "pct_cuda_time": 0.6675825802770947, + "cpu_time_us": 91.476, + "cuda_time_us": 123.413, + "pct_cuda_time": 0.6615834817540126, "trace": "" }, "children": [ @@ -1368,8 +1368,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 118.155, - "pct_cuda_time": 0.6332344800475268, + "cuda_time_us": 116.844, + "pct_cuda_time": 0.6263688618060159, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1378,8 +1378,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.409, - "pct_cuda_time": 0.03434810022956793, + "cuda_time_us": 6.569, + "pct_cuda_time": 0.035214619947996635, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1393,18 +1393,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1089.667, - "cuda_time_us": 570.447, - "pct_cuda_time": 3.0572274507187296, + "cpu_time_us": 1041.097, + "cuda_time_us": 576.657, + "pct_cuda_time": 3.0913011257956917, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 36.244, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cpu_time_us": 35.213, + "cuda_time_us": 9.052, + "pct_cuda_time": 0.048525306708671875, "trace": "" }, "children": [ @@ -1412,8 +1412,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cuda_time_us": 9.052, + "pct_cuda_time": 0.048525306708671875, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1423,18 +1423,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 769.859, - "cuda_time_us": 213.48000000000002, - "pct_cuda_time": 1.1441149066949856, + "cpu_time_us": 722.42, + "cuda_time_us": 213.08, + "pct_cuda_time": 1.1422638481533147, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 86.499, - "cuda_time_us": 54.5, - "pct_cuda_time": 0.29208479677195387, + "cpu_time_us": 85.027, + "cuda_time_us": 55.342, + "pct_cuda_time": 0.2966733897339062, "trace": "" }, "children": [ @@ -1442,8 +1442,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.5, - "pct_cuda_time": 0.29208479677195387, + "cuda_time_us": 55.342, + "pct_cuda_time": 0.2966733897339062, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -1453,9 +1453,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 93.77, - "cuda_time_us": 26.675, - "pct_cuda_time": 0.14296076979618108, + "cpu_time_us": 74.652, + "cuda_time_us": 26.274, + "pct_cuda_time": 0.14084775833668195, "trace": "" }, "children": [ @@ -1463,8 +1463,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -1473,8 +1473,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.405, + "pct_cuda_time": 0.023614005308406943, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -1483,8 +1483,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.729, - "pct_cuda_time": 0.03606309353171518, + "cuda_time_us": 7.01, + "pct_cuda_time": 0.03757870084266348, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -1493,8 +1493,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.727, - "pct_cuda_time": 0.030693020754366602, + "cuda_time_us": 5.767, + "pct_cuda_time": 0.030915316370847407, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -1503,8 +1503,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.087, - "pct_cuda_time": 0.027263034150072092, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -1514,9 +1514,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 445.675, - "cuda_time_us": 89.291, - "pct_cuda_time": 0.47854208419384464, + "cpu_time_us": 428.961, + "cuda_time_us": 90.934, + "pct_cuda_time": 0.48747240833477334, "trace": "" }, "children": [ @@ -1524,8 +1524,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.327, - "pct_cuda_time": 0.028549279126682536, + "cuda_time_us": 5.607, + "pct_cuda_time": 0.030057599946478483, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -1534,8 +1534,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 83.964, - "pct_cuda_time": 0.4499928050671621, + "cuda_time_us": 85.327, + "pct_cuda_time": 0.45741480838829485, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -1545,9 +1545,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 101.682, - "cuda_time_us": 43.014, - "pct_cuda_time": 0.23052725593300594, + "cpu_time_us": 98.287, + "cuda_time_us": 40.53, + "pct_cuda_time": 0.21727029174795306, "trace": "" }, "children": [ @@ -1555,8 +1555,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 43.014, - "pct_cuda_time": 0.23052725593300594, + "cuda_time_us": 40.53, + "pct_cuda_time": 0.21727029174795306, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -1568,9 +1568,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.636, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.05151947066731728, + "cpu_time_us": 38.237, + "cuda_time_us": 9.453, + "pct_cuda_time": 0.05067495849724649, "trace": "" }, "children": [ @@ -1578,8 +1578,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.05151947066731728, + "cuda_time_us": 9.453, + "pct_cuda_time": 0.05067495849724649, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1589,18 +1589,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 224.455, - "cuda_time_us": 338.021, - "pct_cuda_time": 1.8115742218284885, + "cpu_time_us": 224.115, + "cuda_time_us": 345.072, + "pct_cuda_time": 1.8498370124364585, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.738, - "cuda_time_us": 199.657, - "pct_cuda_time": 1.0700325553962935, + "cpu_time_us": 84.346, + "cuda_time_us": 206.798, + "pct_cuda_time": 1.1085877570415297, "trace": "" }, "children": [ @@ -1608,8 +1608,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 199.657, - "pct_cuda_time": 1.0700325553962935, + "cuda_time_us": 206.798, + "pct_cuda_time": 1.1085877570415297, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -1619,9 +1619,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 25.949, - "cuda_time_us": 15.252, - "pct_cuda_time": 0.0817408682635934, + "cpu_time_us": 26.529, + "cuda_time_us": 14.871, + "pct_cuda_time": 0.07971938091743919, "trace": "" }, "children": [ @@ -1629,8 +1629,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.252, - "pct_cuda_time": 0.0817408682635934, + "cuda_time_us": 14.871, + "pct_cuda_time": 0.07971938091743919, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -1640,9 +1640,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 92.848, - "cuda_time_us": 123.11200000000001, - "pct_cuda_time": 0.6598007981686016, + "cpu_time_us": 94.691, + "cuda_time_us": 123.403, + "pct_cuda_time": 0.6615298744774896, "trace": "" }, "children": [ @@ -1650,8 +1650,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 116.983, - "pct_cuda_time": 0.6269533170784125, + "cuda_time_us": 117.194, + "pct_cuda_time": 0.628245116484323, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1660,8 +1660,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.129, - "pct_cuda_time": 0.03284748109018908, + "cuda_time_us": 6.209, + "pct_cuda_time": 0.033284757993166554, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1675,18 +1675,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1053.903, - "cuda_time_us": 575.104, - "pct_cuda_time": 3.0821859626190418, + "cpu_time_us": 1089.248, + "cuda_time_us": 584.239, + "pct_cuda_time": 3.131946162855474, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.001, - "cuda_time_us": 9.292, - "pct_cuda_time": 0.04979911801110083, + "cpu_time_us": 34.441, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "" }, "children": [ @@ -1694,8 +1694,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.292, - "pct_cuda_time": 0.04979911801110083, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1705,18 +1705,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 737.821, - "cuda_time_us": 217.247, - "pct_cuda_time": 1.1643035934737003, + "cpu_time_us": 763.892, + "cuda_time_us": 214.793, + "pct_cuda_time": 1.1514467746217143, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 83.494, - "cuda_time_us": 54.371, - "pct_cuda_time": 0.2913934400970257, + "cpu_time_us": 96.864, + "cuda_time_us": 57.404, + "pct_cuda_time": 0.30772721015296073, "trace": "" }, "children": [ @@ -1724,8 +1724,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.371, - "pct_cuda_time": 0.2913934400970257, + "cuda_time_us": 57.404, + "pct_cuda_time": 0.30772721015296073, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -1735,9 +1735,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 76.183, - "cuda_time_us": 26.634, - "pct_cuda_time": 0.14274103627934348, + "cpu_time_us": 78.307, + "cuda_time_us": 26.715000000000003, + "pct_cuda_time": 0.1432118392313488, "trace": "" }, "children": [ @@ -1745,8 +1745,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.02490594067261263, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -1755,8 +1755,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.485, - "pct_cuda_time": 0.02403670300040758, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -1765,8 +1765,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.33, - "pct_cuda_time": 0.03928406532731049, + "cuda_time_us": 7.611, + "pct_cuda_time": 0.040800498161699256, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -1775,8 +1775,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.447, - "pct_cuda_time": 0.029192401614987755, + "cuda_time_us": 5.367, + "pct_cuda_time": 0.028771025309925098, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -1785,8 +1785,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.565, + "pct_cuda_time": 0.024471721732775863, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -1796,9 +1796,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 439.896, - "cuda_time_us": 90.935, - "pct_cuda_time": 0.4873528622836261, + "cpu_time_us": 450.413, + "cuda_time_us": 91.13499999999999, + "pct_cuda_time": 0.48854991459288677, "trace": "" }, "children": [ @@ -1806,8 +1806,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.608, - "pct_cuda_time": 0.03005525762013059, + "cuda_time_us": 5.648, + "pct_cuda_time": 0.030277389780223016, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -1816,8 +1816,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 85.327, - "pct_cuda_time": 0.4572976046634955, + "cuda_time_us": 85.487, + "pct_cuda_time": 0.45827252481266373, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -1827,9 +1827,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 101.06, - "cuda_time_us": 45.307, - "pct_cuda_time": 0.24281625481370486, + "cpu_time_us": 100.501, + "cuda_time_us": 39.539, + "pct_cuda_time": 0.21195781064451805, "trace": "" }, "children": [ @@ -1837,8 +1837,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 45.307, - "pct_cuda_time": 0.24281625481370486, + "cuda_time_us": 39.539, + "pct_cuda_time": 0.21195781064451805, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -1850,9 +1850,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.797, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cpu_time_us": 38.318, + "cuda_time_us": 9.292, + "pct_cuda_time": 0.04981188134522526, "trace": "" }, "children": [ @@ -1860,8 +1860,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cuda_time_us": 9.292, + "pct_cuda_time": 0.04981188134522526, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1871,18 +1871,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 220.229, - "cuda_time_us": 339.23199999999997, - "pct_cuda_time": 1.8180643996063017, + "cpu_time_us": 229.814, + "cuda_time_us": 350.741, + "pct_cuda_time": 1.8802269774973797, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.217, - "cuda_time_us": 198.856, - "pct_cuda_time": 1.065739712786856, + "cpu_time_us": 87.03, + "cuda_time_us": 209.202, + "pct_cuda_time": 1.1214749463176727, "trace": "" }, "children": [ @@ -1890,8 +1890,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 198.856, - "pct_cuda_time": 1.065739712786856, + "cuda_time_us": 209.202, + "pct_cuda_time": 1.1214749463176727, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -1901,9 +1901,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.459, - "cuda_time_us": 14.931, - "pct_cuda_time": 0.08002051560737693, + "cpu_time_us": 26.55, + "cuda_time_us": 16.063, + "pct_cuda_time": 0.08610936827898766, "trace": "" }, "children": [ @@ -1911,8 +1911,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 14.931, - "pct_cuda_time": 0.08002051560737693, + "cuda_time_us": 16.063, + "pct_cuda_time": 0.08610936827898766, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -1922,9 +1922,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.635, - "cuda_time_us": 125.445, - "pct_cuda_time": 0.6723041712120688, + "cpu_time_us": 93.77, + "cuda_time_us": 125.476, + "pct_cuda_time": 0.6726426629007194, "trace": "" }, "children": [ @@ -1932,8 +1932,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 119.077, - "pct_cuda_time": 0.6381758044993385, + "cuda_time_us": 118.947, + "pct_cuda_time": 0.637642472058815, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1942,8 +1942,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.368, - "pct_cuda_time": 0.03412836671273032, + "cuda_time_us": 6.529, + "pct_cuda_time": 0.03500019084190441, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -1957,18 +1957,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1045.631, - "cuda_time_us": 580.3309999999999, - "pct_cuda_time": 3.110199306338802, + "cpu_time_us": 1051.993, + "cuda_time_us": 583.326, + "pct_cuda_time": 3.1270518185089182, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.991, - "cuda_time_us": 10.255, - "pct_cuda_time": 0.05496017597975022, + "cpu_time_us": 34.321, + "cuda_time_us": 9.293, + "pct_cuda_time": 0.04981724207287757, "trace": "" }, "children": [ @@ -1976,8 +1976,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.255, - "pct_cuda_time": 0.05496017597975022, + "cuda_time_us": 9.293, + "pct_cuda_time": 0.04981724207287757, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -1987,18 +1987,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 722.569, - "cuda_time_us": 216.13299999999998, - "pct_cuda_time": 1.1583332730406, + "cpu_time_us": 731.995, + "cuda_time_us": 219.618, + "pct_cuda_time": 1.1773122855440898, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 86.979, - "cuda_time_us": 57.504, - "pct_cuda_time": 0.30818429639586115, + "cpu_time_us": 83.295, + "cuda_time_us": 54.941, + "pct_cuda_time": 0.29452373794533165, "trace": "" }, "children": [ @@ -2006,8 +2006,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 57.504, - "pct_cuda_time": 0.30818429639586115, + "cuda_time_us": 54.941, + "pct_cuda_time": 0.29452373794533165, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -2017,9 +2017,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.349, - "cuda_time_us": 25.954, - "pct_cuda_time": 0.13909667551228055, + "cpu_time_us": 77.205, + "cuda_time_us": 26.835, + "pct_cuda_time": 0.14385512654962548, "trace": "" }, "children": [ @@ -2027,8 +2027,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -2037,8 +2037,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.446, - "pct_cuda_time": 0.02382768819170838, + "cuda_time_us": 4.405, + "pct_cuda_time": 0.023614005308406943, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -2047,8 +2047,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.328, - "pct_cuda_time": 0.033913992549961906, + "cuda_time_us": 7.53, + "pct_cuda_time": 0.04036627922186249, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -2057,8 +2057,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.928, - "pct_cuda_time": 0.031770250922277844, + "cuda_time_us": 5.848, + "pct_cuda_time": 0.03134953531068417, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -2067,8 +2067,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -2078,9 +2078,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 427.388, - "cuda_time_us": 89.692, - "pct_cuda_time": 0.4806911851755979, + "cpu_time_us": 435.271, + "cuda_time_us": 95.26, + "pct_cuda_time": 0.5106629161586482, "trace": "" }, "children": [ @@ -2088,8 +2088,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.327, - "pct_cuda_time": 0.028549279126682536, + "cuda_time_us": 6.128, + "pct_cuda_time": 0.03285053905332979, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2098,8 +2098,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 84.365, - "pct_cuda_time": 0.4521419060489153, + "cuda_time_us": 89.132, + "pct_cuda_time": 0.4778123771053184, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2109,9 +2109,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 98.958, - "cuda_time_us": 42.983, - "pct_cuda_time": 0.2303611159568604, + "cpu_time_us": 99.739, + "cuda_time_us": 42.582, + "pct_cuda_time": 0.22827050489048453, "trace": "" }, "children": [ @@ -2119,8 +2119,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.983, - "pct_cuda_time": 0.2303611159568604, + "cuda_time_us": 42.582, + "pct_cuda_time": 0.22827050489048453, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -2132,9 +2132,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.88, - "cuda_time_us": 10.414, - "pct_cuda_time": 0.05581231327675463, + "cpu_time_us": 37.887, + "cuda_time_us": 9.574, + "pct_cuda_time": 0.051323606543175485, "trace": "" }, "children": [ @@ -2142,8 +2142,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.414, - "pct_cuda_time": 0.05581231327675463, + "cuda_time_us": 9.574, + "pct_cuda_time": 0.051323606543175485, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2153,18 +2153,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 221.811, - "cuda_time_us": 343.529, - "pct_cuda_time": 1.841093544041698, + "cpu_time_us": 221.161, + "cuda_time_us": 344.841, + "pct_cuda_time": 1.8485986843487758, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 86.389, - "cuda_time_us": 202.692, - "pct_cuda_time": 1.0862981949963464, + "cpu_time_us": 86.289, + "cuda_time_us": 205.977, + "pct_cuda_time": 1.1041865996389866, "trace": "" }, "children": [ @@ -2172,8 +2172,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 202.692, - "pct_cuda_time": 1.0862981949963464, + "cuda_time_us": 205.977, + "pct_cuda_time": 1.1041865996389866, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -2183,9 +2183,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 25.798, - "cuda_time_us": 15.402, - "pct_cuda_time": 0.08254477137397491, + "cpu_time_us": 27.231, + "cuda_time_us": 15.251, + "pct_cuda_time": 0.08175645742531537, "trace": "" }, "children": [ @@ -2193,8 +2193,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.402, - "pct_cuda_time": 0.08254477137397491, + "cuda_time_us": 15.251, + "pct_cuda_time": 0.08175645742531537, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -2204,9 +2204,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.756, - "cuda_time_us": 125.435, - "pct_cuda_time": 0.6722505776713767, + "cpu_time_us": 89.513, + "cuda_time_us": 123.613, + "pct_cuda_time": 0.6626556272844738, "trace": "" }, "children": [ @@ -2214,8 +2214,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 119.307, - "pct_cuda_time": 0.6394084559352569, + "cuda_time_us": 117.084, + "pct_cuda_time": 0.6276554364425694, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -2224,8 +2224,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.128, - "pct_cuda_time": 0.032842121736119875, + "cuda_time_us": 6.529, + "pct_cuda_time": 0.03500019084190441, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -2239,18 +2239,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1040.864, - "cuda_time_us": 581.314, - "pct_cuda_time": 3.115467551388836, + "cpu_time_us": 1052.393, + "cuda_time_us": 579.362, + "pct_cuda_time": 3.1058018940951784, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.57, - "cuda_time_us": 9.734, - "pct_cuda_time": 0.052167952509691716, + "cpu_time_us": 33.23, + "cuda_time_us": 9.613, + "pct_cuda_time": 0.051532674921615416, "trace": "" }, "children": [ @@ -2258,8 +2258,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.734, - "pct_cuda_time": 0.052167952509691716, + "cuda_time_us": 9.613, + "pct_cuda_time": 0.051532674921615416, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2269,18 +2269,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 717.18, - "cuda_time_us": 217.195, - "pct_cuda_time": 1.1640249070621012, + "cpu_time_us": 726.236, + "cuda_time_us": 216.80599999999998, + "pct_cuda_time": 1.162237919385806, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.654, - "cuda_time_us": 57.404, - "pct_cuda_time": 0.30764836098894016, + "cpu_time_us": 85.378, + "cuda_time_us": 57.706, + "pct_cuda_time": 0.3093461499039571, "trace": "" }, "children": [ @@ -2288,8 +2288,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 57.404, - "pct_cuda_time": 0.30764836098894016, + "cuda_time_us": 57.706, + "pct_cuda_time": 0.3093461499039571, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -2299,9 +2299,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 72.178, - "cuda_time_us": 25.994, - "pct_cuda_time": 0.13931104967504895, + "cpu_time_us": 75.363, + "cuda_time_us": 25.953999999999997, + "pct_cuda_time": 0.13913232548794408, "trace": "" }, "children": [ @@ -2309,8 +2309,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.485, + "pct_cuda_time": 0.024042863520591404, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -2319,8 +2319,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.446, - "pct_cuda_time": 0.02382768819170838, + "cuda_time_us": 4.406, + "pct_cuda_time": 0.02361936603605924, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -2329,8 +2329,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.609, - "pct_cuda_time": 0.035419971043409965, + "cuda_time_us": 6.809, + "pct_cuda_time": 0.03650119458455002, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -2339,8 +2339,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.607, - "pct_cuda_time": 0.03004989826606138, + "cuda_time_us": 5.728, + "pct_cuda_time": 0.03070624799240748, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -2349,8 +2349,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -2360,9 +2360,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 427.338, - "cuda_time_us": 90.213, - "pct_cuda_time": 0.48348340864565637, + "cpu_time_us": 430.443, + "cuda_time_us": 91.094, + "pct_cuda_time": 0.4883301247591423, "trace": "" }, "children": [ @@ -2370,8 +2370,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.407, - "pct_cuda_time": 0.02897802745221935, + "cuda_time_us": 5.567, + "pct_cuda_time": 0.029843170840386252, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2380,8 +2380,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 84.806, - "pct_cuda_time": 0.454505381193437, + "cuda_time_us": 85.527, + "pct_cuda_time": 0.458486953918756, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2391,9 +2391,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 99.819, - "cuda_time_us": 43.584, - "pct_cuda_time": 0.23358208775245576, + "cpu_time_us": 99.288, + "cuda_time_us": 42.052, + "pct_cuda_time": 0.22542931923476245, "trace": "" }, "children": [ @@ -2401,8 +2401,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 43.584, - "pct_cuda_time": 0.23358208775245576, + "cuda_time_us": 42.052, + "pct_cuda_time": 0.22542931923476245, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -2414,9 +2414,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.476, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cpu_time_us": 38.668, + "cuda_time_us": 10.455, + "pct_cuda_time": 0.05604640760485688, "trace": "" }, "children": [ @@ -2424,8 +2424,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cuda_time_us": 10.455, + "pct_cuda_time": 0.05604640760485688, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2435,18 +2435,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 232.157, - "cuda_time_us": 344.53099999999995, - "pct_cuda_time": 1.846463616819046, + "cpu_time_us": 232.337, + "cuda_time_us": 342.488, + "pct_cuda_time": 1.8359848921829003, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 84.716, - "cuda_time_us": 206.047, - "pct_cuda_time": 1.1042788278985463, + "cpu_time_us": 87.081, + "cuda_time_us": 203.844, + "pct_cuda_time": 1.0927521675566183, "trace": "" }, "children": [ @@ -2454,8 +2454,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 206.047, - "pct_cuda_time": 1.1042788278985463, + "cuda_time_us": 203.844, + "pct_cuda_time": 1.0927521675566183, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -2465,9 +2465,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 25.979, - "cuda_time_us": 15.051, - "pct_cuda_time": 0.08066363809568215, + "cpu_time_us": 26.289, + "cuda_time_us": 15.582, + "pct_cuda_time": 0.08353085827822859, "trace": "" }, "children": [ @@ -2475,8 +2475,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.051, - "pct_cuda_time": 0.08066363809568215, + "cuda_time_us": 15.582, + "pct_cuda_time": 0.08353085827822859, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -2486,9 +2486,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 104.536, - "cuda_time_us": 123.43299999999999, - "pct_cuda_time": 0.661521150824818, + "cpu_time_us": 95.763, + "cuda_time_us": 123.06200000000001, + "pct_cuda_time": 0.6597018663480534, "trace": "" }, "children": [ @@ -2496,8 +2496,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 117.505, - "pct_cuda_time": 0.6297508999025401, + "cuda_time_us": 116.293, + "pct_cuda_time": 0.6234151008695956, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -2506,8 +2506,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 5.928, - "pct_cuda_time": 0.031770250922277844, + "cuda_time_us": 6.769, + "pct_cuda_time": 0.03628676547845779, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -2521,18 +2521,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1068.154, - "cuda_time_us": 568.855, - "pct_cuda_time": 3.048695359040547, + "cpu_time_us": 1084.211, + "cuda_time_us": 573.752, + "pct_cuda_time": 3.075728211965743, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.436, - "cuda_time_us": 9.573, - "pct_cuda_time": 0.05130509650454888, + "cpu_time_us": 34.972, + "cuda_time_us": 10.535, + "pct_cuda_time": 0.05647526581704134, "trace": "" }, "children": [ @@ -2540,8 +2540,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.573, - "pct_cuda_time": 0.05130509650454888, + "cuda_time_us": 10.535, + "pct_cuda_time": 0.05647526581704134, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2551,18 +2551,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 744.201, - "cuda_time_us": 211.80800000000002, - "pct_cuda_time": 1.1351540666912663, + "cpu_time_us": 744.343, + "cuda_time_us": 211.66699999999997, + "pct_cuda_time": 1.1346891399806065, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 90.996, - "cuda_time_us": 54.5, - "pct_cuda_time": 0.29208479677195387, + "cpu_time_us": 86.519, + "cuda_time_us": 54.72, + "pct_cuda_time": 0.293339017134172, "trace": "" }, "children": [ @@ -2570,8 +2570,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.5, - "pct_cuda_time": 0.29208479677195387, + "cuda_time_us": 54.72, + "pct_cuda_time": 0.293339017134172, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -2581,9 +2581,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 76.324, - "cuda_time_us": 25.754, - "pct_cuda_time": 0.13802480469843853, + "cpu_time_us": 77.125, + "cuda_time_us": 26.796, + "pct_cuda_time": 0.14364605817118553, "trace": "" }, "children": [ @@ -2591,8 +2591,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -2601,8 +2601,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.446, - "pct_cuda_time": 0.02382768819170838, + "cuda_time_us": 4.406, + "pct_cuda_time": 0.02361936603605924, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -2611,8 +2611,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.408, - "pct_cuda_time": 0.03434274087549872, + "cuda_time_us": 7.41, + "pct_cuda_time": 0.039722991903585796, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -2621,8 +2621,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.648, - "pct_cuda_time": 0.030269631782898994, + "cuda_time_us": 5.848, + "pct_cuda_time": 0.03134953531068417, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -2631,8 +2631,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.0246915115665204, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -2642,9 +2642,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 438.765, - "cuda_time_us": 90.173, - "pct_cuda_time": 0.48326903448288794, + "cpu_time_us": 442.522, + "cuda_time_us": 90.493, + "pct_cuda_time": 0.4851083274401065, "trace": "" }, "children": [ @@ -2652,8 +2652,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.327, - "pct_cuda_time": 0.028549279126682536, + "cuda_time_us": 5.487, + "pct_cuda_time": 0.02941431262820179, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2662,8 +2662,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 84.846, - "pct_cuda_time": 0.4547197553562055, + "cuda_time_us": 85.006, + "pct_cuda_time": 0.45569401481190475, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2673,9 +2673,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 102.103, - "cuda_time_us": 41.381, - "pct_cuda_time": 0.22177543073798575, + "cpu_time_us": 101.712, + "cuda_time_us": 39.658, + "pct_cuda_time": 0.21259573723514244, "trace": "" }, "children": [ @@ -2683,8 +2683,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 41.381, - "pct_cuda_time": 0.22177543073798575, + "cuda_time_us": 39.658, + "pct_cuda_time": 0.21259573723514244, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -2696,9 +2696,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.689, - "cuda_time_us": 9.413, - "pct_cuda_time": 0.05044759985347526, + "cpu_time_us": 38.007, + "cuda_time_us": 9.373, + "pct_cuda_time": 0.05024610028506203, "trace": "" }, "children": [ @@ -2706,8 +2706,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.413, - "pct_cuda_time": 0.05044759985347526, + "cuda_time_us": 9.373, + "pct_cuda_time": 0.05024610028506203, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2717,18 +2717,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 222.983, - "cuda_time_us": 338.061, - "pct_cuda_time": 1.8117885959912567, + "cpu_time_us": 242.543, + "cuda_time_us": 342.177, + "pct_cuda_time": 1.8343177058830336, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 86.239, - "cuda_time_us": 198.976, - "pct_cuda_time": 1.0663828352751612, + "cpu_time_us": 98.707, + "cuda_time_us": 200.589, + "pct_cuda_time": 1.0753029990483631, "trace": "" }, "children": [ @@ -2736,8 +2736,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 198.976, - "pct_cuda_time": 1.0663828352751612, + "cuda_time_us": 200.589, + "pct_cuda_time": 1.0753029990483631, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -2747,9 +2747,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.72, - "cuda_time_us": 15.052, - "pct_cuda_time": 0.08066899744975135, + "cpu_time_us": 32.509, + "cuda_time_us": 15.983, + "pct_cuda_time": 0.0856805100668032, "trace": "" }, "children": [ @@ -2757,8 +2757,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.052, - "pct_cuda_time": 0.08066899744975135, + "cuda_time_us": 15.983, + "pct_cuda_time": 0.0856805100668032, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -2768,9 +2768,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 92.589, - "cuda_time_us": 124.033, - "pct_cuda_time": 0.6647367632663441, + "cpu_time_us": 92.238, + "cuda_time_us": 125.60499999999999, + "pct_cuda_time": 0.6733341967678669, "trace": "" }, "children": [ @@ -2778,8 +2778,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 117.224, - "pct_cuda_time": 0.6282449214090922, + "cuda_time_us": 119.237, + "pct_cuda_time": 0.6391970830779836, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -2788,8 +2788,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.809, - "pct_cuda_time": 0.036491841857251996, + "cuda_time_us": 6.368, + "pct_cuda_time": 0.034137113689883175, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -2803,18 +2803,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1044.039, - "cuda_time_us": 586.942, - "pct_cuda_time": 3.145629996090351, + "cpu_time_us": 1059.434, + "cuda_time_us": 581.465, + "pct_cuda_time": 3.117075504347978, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.42, - "cuda_time_us": 9.373, - "pct_cuda_time": 0.05023322569070685, + "cpu_time_us": 33.891, + "cuda_time_us": 9.574, + "pct_cuda_time": 0.051323606543175485, "trace": "" }, "children": [ @@ -2822,8 +2822,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.373, - "pct_cuda_time": 0.05023322569070685, + "cuda_time_us": 9.574, + "pct_cuda_time": 0.051323606543175485, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2833,18 +2833,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 727.837, - "cuda_time_us": 217.16600000000003, - "pct_cuda_time": 1.163869485794094, + "cpu_time_us": 737.002, + "cuda_time_us": 214.34199999999998, + "pct_cuda_time": 1.1490290864505244, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 81.932, - "cuda_time_us": 57.194, - "pct_cuda_time": 0.3065228966344061, + "cpu_time_us": 83.134, + "cuda_time_us": 55.902, + "pct_cuda_time": 0.29967539721919745, "trace": "" }, "children": [ @@ -2852,8 +2852,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 57.194, - "pct_cuda_time": 0.3065228966344061, + "cuda_time_us": 55.902, + "pct_cuda_time": 0.29967539721919745, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -2863,9 +2863,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.523, - "cuda_time_us": 27.116, - "pct_cuda_time": 0.14532424494070276, + "cpu_time_us": 75.142, + "cuda_time_us": 25.512999999999998, + "pct_cuda_time": 0.13676824459327724, "trace": "" }, "children": [ @@ -2874,7 +2874,7 @@ "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -2883,8 +2883,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.446, - "pct_cuda_time": 0.02382768819170838, + "cuda_time_us": 4.366, + "pct_cuda_time": 0.02340493692996701, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -2893,8 +2893,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.169, - "pct_cuda_time": 0.03842120932216765, + "cuda_time_us": 6.448, + "pct_cuda_time": 0.03456597190206764, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -2903,8 +2903,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 6.168, - "pct_cuda_time": 0.03305649589888828, + "cuda_time_us": 5.608, + "pct_cuda_time": 0.030062960674130786, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -2913,8 +2913,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.767, - "pct_cuda_time": 0.025548040847924846, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -2924,9 +2924,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 430.352, - "cuda_time_us": 89.973, - "pct_cuda_time": 0.48219716366904597, + "cpu_time_us": 441.881, + "cuda_time_us": 90.494, + "pct_cuda_time": 0.48511368816775885, "trace": "" }, "children": [ @@ -2934,8 +2934,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.527, - "pct_cuda_time": 0.029621149940524567, + "cuda_time_us": 5.568, + "pct_cuda_time": 0.029848531568038555, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2944,8 +2944,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 84.446, - "pct_cuda_time": 0.45257601372852135, + "cuda_time_us": 84.926, + "pct_cuda_time": 0.4552651565997203, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -2955,9 +2955,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 104.135, - "cuda_time_us": 42.883, - "pct_cuda_time": 0.2298251805499394, + "cpu_time_us": 100.47, + "cuda_time_us": 42.433, + "pct_cuda_time": 0.22747175647029097, "trace": "" }, "children": [ @@ -2965,8 +2965,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.883, - "pct_cuda_time": 0.2298251805499394, + "cuda_time_us": 42.433, + "pct_cuda_time": 0.22747175647029097, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -2978,9 +2978,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.75, - "cuda_time_us": 9.653, - "pct_cuda_time": 0.0517338448300857, + "cpu_time_us": 38.928, + "cuda_time_us": 9.573, + "pct_cuda_time": 0.05131824581552319, "trace": "" }, "children": [ @@ -2988,8 +2988,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.653, - "pct_cuda_time": 0.0517338448300857, + "cuda_time_us": 9.573, + "pct_cuda_time": 0.05131824581552319, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -2999,18 +2999,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 221.42, - "cuda_time_us": 350.75, - "pct_cuda_time": 1.8797934397754643, + "cpu_time_us": 227.08, + "cuda_time_us": 347.976, + "pct_cuda_time": 1.8654045655387543, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 87.961, - "cuda_time_us": 208.641, - "pct_cuda_time": 1.1181809923540775, + "cpu_time_us": 88.612, + "cuda_time_us": 206.518, + "pct_cuda_time": 1.107086753298884, "trace": "" }, "children": [ @@ -3018,8 +3018,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 208.641, - "pct_cuda_time": 1.1181809923540775, + "cuda_time_us": 206.518, + "pct_cuda_time": 1.107086753298884, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -3029,9 +3029,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 25.959, - "cuda_time_us": 16.653, - "pct_cuda_time": 0.08924932331455682, + "cpu_time_us": 27.331, + "cuda_time_us": 15.862, + "pct_cuda_time": 0.0850318620208742, "trace": "" }, "children": [ @@ -3039,8 +3039,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 16.653, - "pct_cuda_time": 0.08924932331455682, + "cuda_time_us": 15.862, + "pct_cuda_time": 0.0850318620208742, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -3050,9 +3050,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.405, - "cuda_time_us": 125.456, - "pct_cuda_time": 0.6723631241068302, + "cpu_time_us": 92.308, + "cuda_time_us": 125.59599999999999, + "pct_cuda_time": 0.6732859502189961, "trace": "" }, "children": [ @@ -3060,8 +3060,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 119.047, - "pct_cuda_time": 0.6380150238772622, + "cuda_time_us": 118.466, + "pct_cuda_time": 0.6350639620580559, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3070,8 +3070,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.409, - "pct_cuda_time": 0.03434810022956793, + "cuda_time_us": 7.13, + "pct_cuda_time": 0.038221988160940176, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3085,18 +3085,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1038.109, - "cuda_time_us": 579.1790000000001, - "pct_cuda_time": 3.104025330451073, + "cpu_time_us": 1065.603, + "cuda_time_us": 576.136, + "pct_cuda_time": 3.08850818668884, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 32.839, - "cuda_time_us": 9.373, - "pct_cuda_time": 0.05023322569070685, + "cpu_time_us": 34.722, + "cuda_time_us": 9.453, + "pct_cuda_time": 0.05067495849724649, "trace": "" }, "children": [ @@ -3104,8 +3104,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.373, - "pct_cuda_time": 0.05023322569070685, + "cuda_time_us": 9.453, + "pct_cuda_time": 0.05067495849724649, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3115,18 +3115,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 728.046, - "cuda_time_us": 216.504, - "pct_cuda_time": 1.1603215934002769, + "cpu_time_us": 745.174, + "cuda_time_us": 216.28400000000002, + "pct_cuda_time": 1.1594396195513024, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.303, - "cuda_time_us": 57.475, - "pct_cuda_time": 0.3080288751278541, + "cpu_time_us": 86.339, + "cuda_time_us": 56.744, + "pct_cuda_time": 0.3041891299024389, "trace": "" }, "children": [ @@ -3134,8 +3134,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 57.475, - "pct_cuda_time": 0.3080288751278541, + "cuda_time_us": 56.744, + "pct_cuda_time": 0.3041891299024389, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -3145,9 +3145,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.433, - "cuda_time_us": 26.354, - "pct_cuda_time": 0.1412404171399646, + "cpu_time_us": 75.503, + "cuda_time_us": 25.513, + "pct_cuda_time": 0.13676824459327724, "trace": "" }, "children": [ @@ -3155,8 +3155,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.605, - "pct_cuda_time": 0.024679825488712802, + "cuda_time_us": 4.486, + "pct_cuda_time": 0.024048224248243707, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -3165,8 +3165,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.406, - "pct_cuda_time": 0.023613314028939975, + "cuda_time_us": 4.445, + "pct_cuda_time": 0.023828434414499174, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -3175,8 +3175,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.21, - "pct_cuda_time": 0.03864094283900527, + "cuda_time_us": 6.369, + "pct_cuda_time": 0.03414247441753548, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -3185,8 +3185,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.527, - "pct_cuda_time": 0.029621149940524567, + "cuda_time_us": 5.647, + "pct_cuda_time": 0.030272029052570714, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -3195,8 +3195,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -3206,9 +3206,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 435.27, - "cuda_time_us": 91.294, - "pct_cuda_time": 0.4892768703944726, + "cpu_time_us": 444.755, + "cuda_time_us": 90.293, + "pct_cuda_time": 0.4840361819096454, "trace": "" }, "children": [ @@ -3216,8 +3216,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.487, - "pct_cuda_time": 0.029406775777756165, + "cuda_time_us": 5.527, + "pct_cuda_time": 0.02962874173429402, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -3226,8 +3226,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 85.807, - "pct_cuda_time": 0.45987009461671646, + "cuda_time_us": 84.766, + "pct_cuda_time": 0.45440744017535134, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -3237,9 +3237,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.29, - "cuda_time_us": 41.381, - "pct_cuda_time": 0.22177543073798575, + "cpu_time_us": 101.853, + "cuda_time_us": 43.734, + "pct_cuda_time": 0.23444606314594077, "trace": "" }, "children": [ @@ -3247,8 +3247,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 41.381, - "pct_cuda_time": 0.22177543073798575, + "cuda_time_us": 43.734, + "pct_cuda_time": 0.23444606314594077, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -3260,9 +3260,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.827, - "cuda_time_us": 9.934, - "pct_cuda_time": 0.053239823323533754, + "cpu_time_us": 40.3, + "cuda_time_us": 9.654, + "pct_cuda_time": 0.05175246475535995, "trace": "" }, "children": [ @@ -3270,8 +3270,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.934, - "pct_cuda_time": 0.053239823323533754, + "cuda_time_us": 9.654, + "pct_cuda_time": 0.05175246475535995, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3281,18 +3281,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 218.436, - "cuda_time_us": 343.36800000000005, - "pct_cuda_time": 1.8402306880365555, + "cpu_time_us": 222.663, + "cuda_time_us": 340.745, + "pct_cuda_time": 1.8266411438849315, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 83.895, - "cuda_time_us": 204.073, - "pct_cuda_time": 1.0936994629659256, + "cpu_time_us": 88.052, + "cuda_time_us": 201.55, + "pct_cuda_time": 1.080454658322229, "trace": "" }, "children": [ @@ -3300,8 +3300,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 204.073, - "pct_cuda_time": 1.0936994629659256, + "cuda_time_us": 201.55, + "pct_cuda_time": 1.080454658322229, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -3311,9 +3311,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.639, - "cuda_time_us": 16.674, - "pct_cuda_time": 0.08936186975001024, + "cpu_time_us": 26.309, + "cuda_time_us": 15.392, + "pct_cuda_time": 0.08251232002429049, "trace": "" }, "children": [ @@ -3321,8 +3321,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 16.674, - "pct_cuda_time": 0.08936186975001024, + "cuda_time_us": 15.392, + "pct_cuda_time": 0.08251232002429049, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -3332,9 +3332,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.805, - "cuda_time_us": 122.62100000000001, - "pct_cuda_time": 0.6571693553206194, + "cpu_time_us": 90.475, + "cuda_time_us": 123.803, + "pct_cuda_time": 0.6636741655384119, "trace": "" }, "children": [ @@ -3342,8 +3342,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 116.373, - "pct_cuda_time": 0.6236841110961943, + "cuda_time_us": 117.996, + "pct_cuda_time": 0.6325444200614723, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3352,8 +3352,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.248, - "pct_cuda_time": 0.033485244224425094, + "cuda_time_us": 5.807, + "pct_cuda_time": 0.031129745476939637, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3367,18 +3367,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1061.555, - "cuda_time_us": 575.774, - "pct_cuda_time": 3.085776729845412, + "cpu_time_us": 1083.86, + "cuda_time_us": 580.6129999999999, + "pct_cuda_time": 3.112508164388213, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 32.889, - "cuda_time_us": 9.413, - "pct_cuda_time": 0.05044759985347526, + "cpu_time_us": 33.531, + "cuda_time_us": 9.694, + "pct_cuda_time": 0.05196689386145219, "trace": "" }, "children": [ @@ -3386,8 +3386,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.413, - "pct_cuda_time": 0.05044759985347526, + "cuda_time_us": 9.694, + "pct_cuda_time": 0.05196689386145219, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3397,18 +3397,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 722.007, - "cuda_time_us": 216.35399999999998, - "pct_cuda_time": 1.1595176902898954, + "cpu_time_us": 746.416, + "cuda_time_us": 221.081, + "pct_cuda_time": 1.185155030099413, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.053, - "cuda_time_us": 57.896, - "pct_cuda_time": 0.3102851631909916, + "cpu_time_us": 90.846, + "cuda_time_us": 57.445, + "pct_cuda_time": 0.30794699998670527, "trace": "" }, "children": [ @@ -3416,8 +3416,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 57.896, - "pct_cuda_time": 0.3102851631909916, + "cuda_time_us": 57.445, + "pct_cuda_time": 0.30794699998670527, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -3427,9 +3427,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 76.013, - "cuda_time_us": 26.113999999999997, - "pct_cuda_time": 0.13995417216335415, + "cpu_time_us": 79.709, + "cuda_time_us": 25.633, + "pct_cuda_time": 0.13741153191155392, "trace": "" }, "children": [ @@ -3437,8 +3437,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.525, - "pct_cuda_time": 0.02425107716317599, + "cuda_time_us": 4.485, + "pct_cuda_time": 0.024042863520591404, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -3447,8 +3447,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.366, - "pct_cuda_time": 0.02339893986617157, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.023833795142151473, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -3457,8 +3457,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.969, - "pct_cuda_time": 0.03734933850832562, + "cuda_time_us": 6.889, + "pct_cuda_time": 0.03693005279673448, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -3467,8 +3467,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.488, - "pct_cuda_time": 0.029412135131825373, + "cuda_time_us": 5.287, + "pct_cuda_time": 0.028342167097740636, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -3477,8 +3477,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.024262653354335938, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -3488,9 +3488,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 428.74, - "cuda_time_us": 92.215, - "pct_cuda_time": 0.4942128354922151, + "cpu_time_us": 438.316, + "cuda_time_us": 97.624, + "pct_cuda_time": 0.5233356763286989, "trace": "" }, "children": [ @@ -3498,8 +3498,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.687, - "pct_cuda_time": 0.030478646591598192, + "cuda_time_us": 5.808, + "pct_cuda_time": 0.031135106204591943, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -3508,8 +3508,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 86.528, - "pct_cuda_time": 0.46373418890061696, + "cuda_time_us": 91.816, + "pct_cuda_time": 0.49220057012410706, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -3519,9 +3519,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.36, - "cuda_time_us": 40.129, - "pct_cuda_time": 0.21506551944333463, + "cpu_time_us": 100.871, + "cuda_time_us": 40.379, + "pct_cuda_time": 0.2164608218724549, "trace": "" }, "children": [ @@ -3529,8 +3529,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 40.129, - "pct_cuda_time": 0.21506551944333463, + "cuda_time_us": 40.379, + "pct_cuda_time": 0.2164608218724549, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -3542,9 +3542,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 60.14, - "cuda_time_us": 9.693, - "pct_cuda_time": 0.0519482189928541, + "cpu_time_us": 47.1, + "cuda_time_us": 10.054, + "pct_cuda_time": 0.05389675581628226, "trace": "" }, "children": [ @@ -3552,8 +3552,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.693, - "pct_cuda_time": 0.0519482189928541, + "cuda_time_us": 10.054, + "pct_cuda_time": 0.05389675581628226, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3563,18 +3563,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 225.656, - "cuda_time_us": 340.314, - "pct_cuda_time": 1.8238632207091872, + "cpu_time_us": 233.93, + "cuda_time_us": 339.784, + "pct_cuda_time": 1.8214894846110656, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 88.522, - "cuda_time_us": 200.028, - "pct_cuda_time": 1.0720208757559702, + "cpu_time_us": 91.848, + "cuda_time_us": 201.009, + "pct_cuda_time": 1.0775545046623316, "trace": "" }, "children": [ @@ -3582,8 +3582,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 200.028, - "pct_cuda_time": 1.0720208757559702, + "cuda_time_us": 201.009, + "pct_cuda_time": 1.0775545046623316, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -3593,9 +3593,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.71, - "cuda_time_us": 15.252, - "pct_cuda_time": 0.0817408682635934, + "cpu_time_us": 29.274, + "cuda_time_us": 14.421, + "pct_cuda_time": 0.07730705347390159, "trace": "" }, "children": [ @@ -3603,8 +3603,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.252, - "pct_cuda_time": 0.0817408682635934, + "cuda_time_us": 14.421, + "pct_cuda_time": 0.07730705347390159, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -3614,9 +3614,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 92.368, - "cuda_time_us": 125.034, - "pct_cuda_time": 0.6701014766896235, + "cpu_time_us": 93.73, + "cuda_time_us": 124.354, + "pct_cuda_time": 0.6666279264748324, "trace": "" }, "children": [ @@ -3624,8 +3624,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 118.786, - "pct_cuda_time": 0.6366162324651984, + "cuda_time_us": 117.585, + "pct_cuda_time": 0.6303411609963745, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3634,8 +3634,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.248, - "pct_cuda_time": 0.033485244224425094, + "cuda_time_us": 6.769, + "pct_cuda_time": 0.03628676547845779, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3649,18 +3649,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1057.519, - "cuda_time_us": 577.828, - "pct_cuda_time": 3.0967848431035696, + "cpu_time_us": 1095.779, + "cuda_time_us": 585.4200000000001, + "pct_cuda_time": 3.1382771822128475, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.301, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cpu_time_us": 34.772, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "" }, "children": [ @@ -3668,8 +3668,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3679,18 +3679,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 738.522, - "cuda_time_us": 212.658, - "pct_cuda_time": 1.1397095176500947, + "cpu_time_us": 769.16, + "cuda_time_us": 220.06, + "pct_cuda_time": 1.179681727166409, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 84.075, - "cuda_time_us": 54.951, - "pct_cuda_time": 0.29450186545716767, + "cpu_time_us": 85.447, + "cuda_time_us": 57.264, + "pct_cuda_time": 0.3069767082816379, "trace": "" }, "children": [ @@ -3698,8 +3698,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.951, - "pct_cuda_time": 0.29450186545716767, + "cuda_time_us": 57.264, + "pct_cuda_time": 0.3069767082816379, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -3709,9 +3709,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 74.772, - "cuda_time_us": 26.394999999999996, - "pct_cuda_time": 0.14146015065680223, + "cpu_time_us": 77.656, + "cuda_time_us": 25.433, + "pct_cuda_time": 0.13633938638109278, "trace": "" }, "children": [ @@ -3719,8 +3719,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.486, + "pct_cuda_time": 0.024048224248243707, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -3729,8 +3729,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.366, - "pct_cuda_time": 0.02339893986617157, + "cuda_time_us": 4.486, + "pct_cuda_time": 0.024048224248243707, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -3739,8 +3739,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.29, - "pct_cuda_time": 0.039069691164542085, + "cuda_time_us": 6.609, + "pct_cuda_time": 0.03542904905408887, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -3749,8 +3749,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.447, - "pct_cuda_time": 0.029192401614987755, + "cuda_time_us": 5.327, + "pct_cuda_time": 0.028556596203832867, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -3759,8 +3759,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -3770,9 +3770,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 430.633, - "cuda_time_us": 90.813, - "pct_cuda_time": 0.4866990210871825, + "cpu_time_us": 458.475, + "cuda_time_us": 95.421, + "pct_cuda_time": 0.5115259933106694, "trace": "" }, "children": [ @@ -3780,8 +3780,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.367, - "pct_cuda_time": 0.028763653289450943, + "cuda_time_us": 5.888, + "pct_cuda_time": 0.031563964416776405, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -3790,8 +3790,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 85.446, - "pct_cuda_time": 0.45793536779773153, + "cuda_time_us": 89.533, + "pct_cuda_time": 0.479962028893893, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -3801,9 +3801,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 108.011, - "cuda_time_us": 40.499, - "pct_cuda_time": 0.21704848044894237, + "cpu_time_us": 106.439, + "cuda_time_us": 41.942, + "pct_cuda_time": 0.22483963919300884, "trace": "" }, "children": [ @@ -3811,8 +3811,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 40.499, - "pct_cuda_time": 0.21704848044894237, + "cuda_time_us": 41.942, + "pct_cuda_time": 0.22483963919300884, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -3824,9 +3824,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.507, - "cuda_time_us": 9.494, - "pct_cuda_time": 0.050881707533081286, + "cpu_time_us": 39.609, + "cuda_time_us": 9.854, + "pct_cuda_time": 0.0528246102858211, "trace": "" }, "children": [ @@ -3834,8 +3834,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.494, - "pct_cuda_time": 0.050881707533081286, + "cuda_time_us": 9.854, + "pct_cuda_time": 0.0528246102858211, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3845,18 +3845,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 223.213, - "cuda_time_us": 346.343, - "pct_cuda_time": 1.8561747663924555, + "cpu_time_us": 229.773, + "cuda_time_us": 345.572, + "pct_cuda_time": 1.8525173762626115, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 87.711, - "cuda_time_us": 205.536, - "pct_cuda_time": 1.10154019796918, + "cpu_time_us": 90.215, + "cuda_time_us": 206.267, + "pct_cuda_time": 1.1057412106581554, "trace": "" }, "children": [ @@ -3864,8 +3864,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 205.536, - "pct_cuda_time": 1.10154019796918, + "cuda_time_us": 206.267, + "pct_cuda_time": 1.1057412106581554, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -3875,9 +3875,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.559, - "cuda_time_us": 15.722, - "pct_cuda_time": 0.08425976467612217, + "cpu_time_us": 27.892, + "cuda_time_us": 15.973, + "pct_cuda_time": 0.08562690279028015, "trace": "" }, "children": [ @@ -3885,8 +3885,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.722, - "pct_cuda_time": 0.08425976467612217, + "cuda_time_us": 15.973, + "pct_cuda_time": 0.08562690279028015, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -3896,9 +3896,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.736, - "cuda_time_us": 125.085, - "pct_cuda_time": 0.6703748037471532, + "cpu_time_us": 92.478, + "cuda_time_us": 123.33200000000001, + "pct_cuda_time": 0.661149262814176, "trace": "" }, "children": [ @@ -3906,8 +3906,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 118.356, - "pct_cuda_time": 0.634311710215438, + "cuda_time_us": 117.525, + "pct_cuda_time": 0.6300195173372363, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3916,8 +3916,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.729, - "pct_cuda_time": 0.03606309353171518, + "cuda_time_us": 5.807, + "pct_cuda_time": 0.031129745476939637, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -3931,18 +3931,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1042.326, - "cuda_time_us": 572.538, - "pct_cuda_time": 3.068433860077448, + "cpu_time_us": 1076.479, + "cuda_time_us": 568.836, + "pct_cuda_time": 3.049374874827008, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.941, - "cuda_time_us": 9.253, - "pct_cuda_time": 0.049590103202401634, + "cpu_time_us": 37.225, + "cuda_time_us": 9.373, + "pct_cuda_time": 0.05024610028506203, "trace": "" }, "children": [ @@ -3950,8 +3950,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.253, - "pct_cuda_time": 0.049590103202401634, + "cuda_time_us": 9.373, + "pct_cuda_time": 0.05024610028506203, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -3961,18 +3961,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 727.496, - "cuda_time_us": 215.862, - "pct_cuda_time": 1.156880888087844, + "cpu_time_us": 749.481, + "cuda_time_us": 212.83999999999997, + "pct_cuda_time": 1.1409772735167611, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 83.515, - "cuda_time_us": 55.411, - "pct_cuda_time": 0.2969671683290043, + "cpu_time_us": 86.229, + "cuda_time_us": 56.514, + "pct_cuda_time": 0.3029561625424086, "trace": "" }, "children": [ @@ -3980,8 +3980,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 55.411, - "pct_cuda_time": 0.2969671683290043, + "cuda_time_us": 56.514, + "pct_cuda_time": 0.3029561625424086, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -3991,9 +3991,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.513, - "cuda_time_us": 25.994, - "pct_cuda_time": 0.13931104967504895, + "cpu_time_us": 78.677, + "cuda_time_us": 25.713, + "pct_cuda_time": 0.13784039012373842, "trace": "" }, "children": [ @@ -4001,8 +4001,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.024899559005550416, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -4011,8 +4011,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.405, + "pct_cuda_time": 0.023614005308406943, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -4021,8 +4021,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.689, - "pct_cuda_time": 0.03584871936894678, + "cuda_time_us": 6.609, + "pct_cuda_time": 0.03542904905408887, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -4031,8 +4031,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.287, - "pct_cuda_time": 0.028334904963914127, + "cuda_time_us": 5.447, + "pct_cuda_time": 0.02919988352210956, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -4041,8 +4041,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.02597142981939245, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -4052,9 +4052,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 433.878, - "cuda_time_us": 91.574, - "pct_cuda_time": 0.4907774895338514, + "cpu_time_us": 445.396, + "cuda_time_us": 91.295, + "pct_cuda_time": 0.48940763101725576, "trace": "" }, "children": [ @@ -4062,8 +4062,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.527, - "pct_cuda_time": 0.029621149940524567, + "cuda_time_us": 5.367, + "pct_cuda_time": 0.028771025309925098, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4072,8 +4072,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 86.047, - "pct_cuda_time": 0.46115633959332686, + "cuda_time_us": 85.928, + "pct_cuda_time": 0.46063660570733067, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4083,9 +4083,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 98.998, - "cuda_time_us": 42.883, - "pct_cuda_time": 0.2298251805499394, + "cpu_time_us": 102.654, + "cuda_time_us": 39.318, + "pct_cuda_time": 0.21077308983335846, "trace": "" }, "children": [ @@ -4093,8 +4093,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.883, - "pct_cuda_time": 0.2298251805499394, + "cuda_time_us": 39.318, + "pct_cuda_time": 0.21077308983335846, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -4106,9 +4106,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.259, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.05151947066731728, + "cpu_time_us": 38.197, + "cuda_time_us": 9.693, + "pct_cuda_time": 0.05196153313379988, "trace": "" }, "children": [ @@ -4116,8 +4116,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.05151947066731728, + "cuda_time_us": 9.693, + "pct_cuda_time": 0.05196153313379988, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4127,18 +4127,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 219.888, - "cuda_time_us": 337.81, - "pct_cuda_time": 1.810443398119885, + "cpu_time_us": 225.187, + "cuda_time_us": 336.93, + "pct_cuda_time": 1.8061899678913849, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 84.125, - "cuda_time_us": 200.699, - "pct_cuda_time": 1.0756170023364104, + "cpu_time_us": 85.247, + "cuda_time_us": 197.675, + "pct_cuda_time": 1.0596818386695441, "trace": "" }, "children": [ @@ -4146,8 +4146,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 200.699, - "pct_cuda_time": 1.0756170023364104, + "cuda_time_us": 197.675, + "pct_cuda_time": 1.0596818386695441, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -4157,9 +4157,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.02, - "cuda_time_us": 14.1, - "pct_cuda_time": 0.07556689237586328, + "cpu_time_us": 26.97, + "cuda_time_us": 15.162, + "pct_cuda_time": 0.08127935266426017, "trace": "" }, "children": [ @@ -4167,8 +4167,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 14.1, - "pct_cuda_time": 0.07556689237586328, + "cuda_time_us": 15.162, + "pct_cuda_time": 0.08127935266426017, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -4178,9 +4178,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 91.387, - "cuda_time_us": 123.01100000000001, - "pct_cuda_time": 0.6592595034076113, + "cpu_time_us": 94.471, + "cuda_time_us": 124.093, + "pct_cuda_time": 0.6652287765575806, "trace": "" }, "children": [ @@ -4188,8 +4188,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 116.763, - "pct_cuda_time": 0.6257742591831862, + "cuda_time_us": 117.765, + "pct_cuda_time": 0.6313060919737896, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -4198,8 +4198,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.248, - "pct_cuda_time": 0.033485244224425094, + "cuda_time_us": 6.328, + "pct_cuda_time": 0.03392268458379095, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -4213,18 +4213,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1039.783, - "cuda_time_us": 580.462, - "pct_cuda_time": 3.1109013817218694, + "cpu_time_us": 1072.573, + "cuda_time_us": 580.302, + "pct_cuda_time": 3.1108409780883464, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.46, - "cuda_time_us": 9.894, - "pct_cuda_time": 0.053025449160765355, + "cpu_time_us": 34.241, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "" }, "children": [ @@ -4232,8 +4232,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.894, - "pct_cuda_time": 0.053025449160765355, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4243,18 +4243,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 729.399, - "cuda_time_us": 217.106, - "pct_cuda_time": 1.1635479245499414, + "cpu_time_us": 743.541, + "cuda_time_us": 217.406, + "pct_cuda_time": 1.1654543559771893, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 83.304, - "cuda_time_us": 54.22, - "pct_cuda_time": 0.290584177632575, + "cpu_time_us": 86.339, + "cuda_time_us": 57.165, + "pct_cuda_time": 0.3064459962440596, "trace": "" }, "children": [ @@ -4262,8 +4262,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.22, - "pct_cuda_time": 0.290584177632575, + "cuda_time_us": 57.165, + "pct_cuda_time": 0.3064459962440596, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -4273,9 +4273,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 74.291, - "cuda_time_us": 26.595, - "pct_cuda_time": 0.14253202147064425, + "cpu_time_us": 77.035, + "cuda_time_us": 26.114, + "pct_cuda_time": 0.139990041912313, "trace": "" }, "children": [ @@ -4283,8 +4283,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -4293,8 +4293,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.023833795142151473, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -4303,8 +4303,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.209, - "pct_cuda_time": 0.03863558348493606, + "cuda_time_us": 7.25, + "pct_cuda_time": 0.03886527547921687, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -4313,8 +4313,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.488, - "pct_cuda_time": 0.029412135131825373, + "cuda_time_us": 5.327, + "pct_cuda_time": 0.028556596203832867, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -4323,8 +4323,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -4334,9 +4334,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 433.107, - "cuda_time_us": 91.695, - "pct_cuda_time": 0.49142597137622585, + "cpu_time_us": 442.331, + "cuda_time_us": 91.374, + "pct_cuda_time": 0.4898311285017879, "trace": "" }, "children": [ @@ -4344,8 +4344,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.327, - "pct_cuda_time": 0.028549279126682536, + "cuda_time_us": 5.527, + "pct_cuda_time": 0.02962874173429402, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4354,8 +4354,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 86.368, - "pct_cuda_time": 0.46287669224954325, + "cuda_time_us": 85.847, + "pct_cuda_time": 0.4602023867674938, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4365,9 +4365,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 102.914, - "cuda_time_us": 44.596, - "pct_cuda_time": 0.2390057540704964, + "cpu_time_us": 101.802, + "cuda_time_us": 42.753, + "pct_cuda_time": 0.22918718931902882, "trace": "" }, "children": [ @@ -4375,8 +4375,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 44.596, - "pct_cuda_time": 0.2390057540704964, + "cuda_time_us": 42.753, + "pct_cuda_time": 0.22918718931902882, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -4388,9 +4388,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.088, - "cuda_time_us": 9.573, - "pct_cuda_time": 0.05130509650454888, + "cpu_time_us": 38.067, + "cuda_time_us": 10.415, + "pct_cuda_time": 0.055831978498764645, "trace": "" }, "children": [ @@ -4398,8 +4398,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.573, - "pct_cuda_time": 0.05130509650454888, + "cuda_time_us": 10.415, + "pct_cuda_time": 0.055831978498764645, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4409,18 +4409,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 216.653, - "cuda_time_us": 343.889, - "pct_cuda_time": 1.8430229115066137, + "cpu_time_us": 230.795, + "cuda_time_us": 343.068, + "pct_cuda_time": 1.8390941142212376, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.718, - "cuda_time_us": 205.736, - "pct_cuda_time": 1.1026120687830219, + "cpu_time_us": 90.515, + "cuda_time_us": 203.343, + "pct_cuda_time": 1.0900664430028133, "trace": "" }, "children": [ @@ -4428,8 +4428,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 205.736, - "pct_cuda_time": 1.1026120687830219, + "cuda_time_us": 203.343, + "pct_cuda_time": 1.0900664430028133, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -4439,9 +4439,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.529, - "cuda_time_us": 15.292, - "pct_cuda_time": 0.0819552424263618, + "cpu_time_us": 27.681, + "cuda_time_us": 16.153, + "pct_cuda_time": 0.08659183376769518, "trace": "" }, "children": [ @@ -4449,8 +4449,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 15.292, - "pct_cuda_time": 0.0819552424263618, + "cuda_time_us": 16.153, + "pct_cuda_time": 0.08659183376769518, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -4460,9 +4460,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 86.87, - "cuda_time_us": 122.861, - "pct_cuda_time": 0.6584556002972298, + "cpu_time_us": 93.009, + "cuda_time_us": 123.572, + "pct_cuda_time": 0.6624358374507293, "trace": "" }, "children": [ @@ -4470,8 +4470,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 116.453, - "pct_cuda_time": 0.6241128594217311, + "cuda_time_us": 117.364, + "pct_cuda_time": 0.629156440185215, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -4480,8 +4480,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.408, - "pct_cuda_time": 0.03434274087549872, + "cuda_time_us": 6.208, + "pct_cuda_time": 0.03327939726551425, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -4495,18 +4495,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1081.094, - "cuda_time_us": 574.472, - "pct_cuda_time": 3.0787988508473, + "cpu_time_us": 1081.527, + "cuda_time_us": 576.446, + "pct_cuda_time": 3.090170012261055, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.36, - "cuda_time_us": 9.774, - "pct_cuda_time": 0.05238232667246012, + "cpu_time_us": 34.692, + "cuda_time_us": 9.133, + "pct_cuda_time": 0.04895952564850864, "trace": "" }, "children": [ @@ -4514,8 +4514,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.774, - "pct_cuda_time": 0.05238232667246012, + "cuda_time_us": 9.133, + "pct_cuda_time": 0.04895952564850864, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4525,18 +4525,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 732.113, - "cuda_time_us": 216.043, - "pct_cuda_time": 1.1578509311743712, + "cpu_time_us": 751.724, + "cuda_time_us": 214.892, + "pct_cuda_time": 1.1519774866592927, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 81.322, - "cuda_time_us": 54.961, - "pct_cuda_time": 0.29455545899785973, + "cpu_time_us": 87.01, + "cuda_time_us": 53.78, + "pct_cuda_time": 0.2882999331410046, "trace": "" }, "children": [ @@ -4544,8 +4544,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.961, - "pct_cuda_time": 0.29455545899785973, + "cuda_time_us": 53.78, + "pct_cuda_time": 0.2882999331410046, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -4555,9 +4555,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.8, - "cuda_time_us": 26.354, - "pct_cuda_time": 0.1412404171399646, + "cpu_time_us": 76.635, + "cuda_time_us": 25.793, + "pct_cuda_time": 0.13826924833592286, "trace": "" }, "children": [ @@ -4565,8 +4565,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -4575,8 +4575,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.405, - "pct_cuda_time": 0.023607954674870767, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.023833795142151473, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -4585,8 +4585,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.01, - "pct_cuda_time": 0.03756907202516324, + "cuda_time_us": 6.769, + "pct_cuda_time": 0.03628676547845779, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -4595,8 +4595,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.687, - "pct_cuda_time": 0.030478646591598192, + "cuda_time_us": 5.447, + "pct_cuda_time": 0.02919988352210956, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -4605,8 +4605,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.0246915115665204, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -4616,9 +4616,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 441.94, - "cuda_time_us": 91.775, - "pct_cuda_time": 0.49185471970176264, + "cpu_time_us": 451.646, + "cuda_time_us": 91.41399999999999, + "pct_cuda_time": 0.49004555760788004, "trace": "" }, "children": [ @@ -4626,8 +4626,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.608, - "pct_cuda_time": 0.03005525762013059, + "cuda_time_us": 5.567, + "pct_cuda_time": 0.029843170840386252, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4636,8 +4636,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 86.167, - "pct_cuda_time": 0.46179946208163203, + "cuda_time_us": 85.847, + "pct_cuda_time": 0.4602023867674938, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4647,9 +4647,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 99.389, - "cuda_time_us": 42.953, - "pct_cuda_time": 0.23020033533478415, + "cpu_time_us": 99.609, + "cuda_time_us": 43.905, + "pct_cuda_time": 0.23536274757448508, "trace": "" }, "children": [ @@ -4657,8 +4657,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.953, - "pct_cuda_time": 0.23020033533478415, + "cuda_time_us": 43.905, + "pct_cuda_time": 0.23536274757448508, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -4670,9 +4670,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 45.808, - "cuda_time_us": 9.573, - "pct_cuda_time": 0.05130509650454888, + "cpu_time_us": 46.729, + "cuda_time_us": 9.613, + "pct_cuda_time": 0.051532674921615416, "trace": "" }, "children": [ @@ -4680,8 +4680,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.573, - "pct_cuda_time": 0.05130509650454888, + "cuda_time_us": 9.613, + "pct_cuda_time": 0.051532674921615416, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4691,18 +4691,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 248.471, - "cuda_time_us": 339.082, - "pct_cuda_time": 1.81726049649592, + "cpu_time_us": 223.514, + "cuda_time_us": 342.808, + "pct_cuda_time": 1.8377003250316382, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 103.855, - "cuda_time_us": 205.416, - "pct_cuda_time": 1.1008970754808747, + "cpu_time_us": 88.943, + "cuda_time_us": 205.186, + "pct_cuda_time": 1.099946264066013, "trace": "" }, "children": [ @@ -4710,8 +4710,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 205.416, - "pct_cuda_time": 1.1008970754808747, + "cuda_time_us": 205.186, + "pct_cuda_time": 1.099946264066013, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -4721,9 +4721,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 34.742, - "cuda_time_us": 13.058, - "pct_cuda_time": 0.0699824454357463, + "cpu_time_us": 28.212, + "cuda_time_us": 14.751, + "pct_cuda_time": 0.07907609359916248, "trace": "" }, "children": [ @@ -4731,8 +4731,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.058, - "pct_cuda_time": 0.0699824454357463, + "cuda_time_us": 14.751, + "pct_cuda_time": 0.07907609359916248, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -4742,9 +4742,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 91.717, - "cuda_time_us": 120.608, - "pct_cuda_time": 0.6463809755792993, + "cpu_time_us": 88.262, + "cuda_time_us": 122.871, + "pct_cuda_time": 0.6586779673664629, "trace": "" }, "children": [ @@ -4752,8 +4752,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 113.839, - "pct_cuda_time": 0.6101035078848157, + "cuda_time_us": 115.862, + "pct_cuda_time": 0.6211046272514517, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -4762,8 +4762,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.769, - "pct_cuda_time": 0.03627746769448359, + "cuda_time_us": 7.009, + "pct_cuda_time": 0.037573340115011185, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -4777,18 +4777,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1048.225, - "cuda_time_us": 565.97, - "pct_cuda_time": 3.033233622550876, + "cpu_time_us": 1083.68, + "cuda_time_us": 563.8689999999999, + "pct_cuda_time": 3.0227481405780043, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.673, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cpu_time_us": 33.12, + "cuda_time_us": 10.055, + "pct_cuda_time": 0.05390211654393457, "trace": "" }, "children": [ @@ -4796,8 +4796,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.05001885152793845, + "cuda_time_us": 10.055, + "pct_cuda_time": 0.05390211654393457, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4807,18 +4807,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 727.205, - "cuda_time_us": 215.573, - "pct_cuda_time": 1.1553320347618423, + "cpu_time_us": 762.019, + "cuda_time_us": 216.305, + "pct_cuda_time": 1.1595521948320007, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 84.636, - "cuda_time_us": 53.509, - "pct_cuda_time": 0.28677367688936656, + "cpu_time_us": 83.745, + "cuda_time_us": 54.62, + "pct_cuda_time": 0.29280294436894144, "trace": "" }, "children": [ @@ -4826,8 +4826,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 53.509, - "pct_cuda_time": 0.28677367688936656, + "cuda_time_us": 54.62, + "pct_cuda_time": 0.29280294436894144, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -4837,9 +4837,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.663, - "cuda_time_us": 26.955, - "pct_cuda_time": 0.1444613889355599, + "cpu_time_us": 73.67, + "cuda_time_us": 26.273999999999997, + "pct_cuda_time": 0.14084775833668192, "trace": "" }, "children": [ @@ -4847,8 +4847,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -4857,8 +4857,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.486, + "pct_cuda_time": 0.024048224248243707, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -4867,8 +4867,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.25, - "pct_cuda_time": 0.03885531700177368, + "cuda_time_us": 6.929, + "pct_cuda_time": 0.037144481902826716, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -4877,8 +4877,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.647, - "pct_cuda_time": 0.03026427242882979, + "cuda_time_us": 5.447, + "pct_cuda_time": 0.02919988352210956, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -4887,8 +4887,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -4898,9 +4898,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 431.063, - "cuda_time_us": 92.216, - "pct_cuda_time": 0.4942181948462843, + "cpu_time_us": 461.3, + "cuda_time_us": 91.055, + "pct_cuda_time": 0.4881210563807024, "trace": "" }, "children": [ @@ -4908,8 +4908,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.247, - "pct_cuda_time": 0.028120530801145724, + "cuda_time_us": 5.408, + "pct_cuda_time": 0.02899081514366963, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4918,8 +4918,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 86.969, - "pct_cuda_time": 0.46609766404513864, + "cuda_time_us": 85.647, + "pct_cuda_time": 0.4591302412370327, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -4929,9 +4929,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 101.281, - "cuda_time_us": 42.893, - "pct_cuda_time": 0.2298787740906315, + "cpu_time_us": 107.351, + "cuda_time_us": 44.356, + "pct_cuda_time": 0.23778043574567498, "trace": "" }, "children": [ @@ -4939,8 +4939,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.893, - "pct_cuda_time": 0.2298787740906315, + "cuda_time_us": 44.356, + "pct_cuda_time": 0.23778043574567498, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -4952,9 +4952,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.917, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cpu_time_us": 37.756, + "cuda_time_us": 9.373, + "pct_cuda_time": 0.05024610028506203, "trace": "" }, "children": [ @@ -4962,8 +4962,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cuda_time_us": 9.373, + "pct_cuda_time": 0.05024610028506203, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -4973,18 +4973,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 225.166, - "cuda_time_us": 331.21000000000004, - "pct_cuda_time": 1.7750716612630981, + "cpu_time_us": 228.171, + "cuda_time_us": 328.13599999999997, + "pct_cuda_time": 1.7590477289170077, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 86.67, - "cuda_time_us": 195.561, - "pct_cuda_time": 1.0480806411288086, + "cpu_time_us": 92.529, + "cuda_time_us": 195.141, + "pct_cuda_time": 1.0460977547986012, "trace": "" }, "children": [ @@ -4992,8 +4992,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 195.561, - "pct_cuda_time": 1.0480806411288086, + "cuda_time_us": 195.141, + "pct_cuda_time": 1.0460977547986012, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -5003,9 +5003,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.051, - "cuda_time_us": 14.18, - "pct_cuda_time": 0.07599564070140011, + "cpu_time_us": 27.071, + "cuda_time_us": 13.459, + "pct_cuda_time": 0.07215003347238343, "trace": "" }, "children": [ @@ -5013,8 +5013,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 14.18, - "pct_cuda_time": 0.07599564070140011, + "cuda_time_us": 13.459, + "pct_cuda_time": 0.07215003347238343, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -5024,9 +5024,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 92.197, - "cuda_time_us": 121.46900000000001, - "pct_cuda_time": 0.6509953794328892, + "cpu_time_us": 90.776, + "cuda_time_us": 119.536, + "pct_cuda_time": 0.6407999406460232, "trace": "" }, "children": [ @@ -5034,8 +5034,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 115.061, - "pct_cuda_time": 0.6166526385573906, + "cuda_time_us": 113.488, + "pct_cuda_time": 0.6083782598048778, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5044,8 +5044,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.408, - "pct_cuda_time": 0.03434274087549872, + "cuda_time_us": 6.048, + "pct_cuda_time": 0.03242168084114533, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5059,18 +5059,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1037.689, - "cuda_time_us": 571.789, - "pct_cuda_time": 3.0644197038796097, + "cpu_time_us": 1088.407, + "cuda_time_us": 577.3689999999999, + "pct_cuda_time": 3.0951179638841326, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.412, - "cuda_time_us": 9.814, - "pct_cuda_time": 0.052596700835228535, + "cpu_time_us": 34.371, + "cuda_time_us": 9.774, + "pct_cuda_time": 0.052395752073636635, "trace": "" }, "children": [ @@ -5078,8 +5078,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.814, - "pct_cuda_time": 0.052596700835228535, + "cuda_time_us": 9.774, + "pct_cuda_time": 0.052395752073636635, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5089,18 +5089,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 725.533, - "cuda_time_us": 218.196, - "pct_cuda_time": 1.1693896204853806, + "cpu_time_us": 765.294, + "cuda_time_us": 220.351, + "pct_cuda_time": 1.1812416989132297, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 84.967, - "cuda_time_us": 54.781, - "pct_cuda_time": 0.2935907752654019, + "cpu_time_us": 83.535, + "cuda_time_us": 53.95, + "pct_cuda_time": 0.2892112568418966, "trace": "" }, "children": [ @@ -5108,8 +5108,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.781, - "pct_cuda_time": 0.2935907752654019, + "cuda_time_us": 53.95, + "pct_cuda_time": 0.2892112568418966, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -5119,9 +5119,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.322, - "cuda_time_us": 26.875, - "pct_cuda_time": 0.1440326406100231, + "cpu_time_us": 83.085, + "cuda_time_us": 26.073999999999998, + "pct_cuda_time": 0.13977561280622078, "trace": "" }, "children": [ @@ -5129,8 +5129,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.605, + "pct_cuda_time": 0.024686150838868093, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -5139,8 +5139,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.023833795142151473, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -5149,8 +5149,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.608, - "pct_cuda_time": 0.03541461168934075, + "cuda_time_us": 7.09, + "pct_cuda_time": 0.03800755905484794, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -5159,8 +5159,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 6.049, - "pct_cuda_time": 0.03241873276465227, + "cuda_time_us": 5.968, + "pct_cuda_time": 0.03199282262896087, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -5169,8 +5169,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.02597142981939245, + "cuda_time_us": 3.965, + "pct_cuda_time": 0.021255285141392397, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -5180,9 +5180,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 431.143, - "cuda_time_us": 93.217, - "pct_cuda_time": 0.4995829082695637, + "cpu_time_us": 460.789, + "cuda_time_us": 98.225, + "pct_cuda_time": 0.5265574736477348, "trace": "" }, "children": [ @@ -5190,8 +5190,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.727, - "pct_cuda_time": 0.030693020754366602, + "cuda_time_us": 6.449, + "pct_cuda_time": 0.03457133262971994, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -5200,8 +5200,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 87.49, - "pct_cuda_time": 0.4688898875151971, + "cuda_time_us": 91.776, + "pct_cuda_time": 0.4919861410180148, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -5211,9 +5211,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 98.087, - "cuda_time_us": 43.323, - "pct_cuda_time": 0.23218329634039186, + "cpu_time_us": 102.864, + "cuda_time_us": 42.102, + "pct_cuda_time": 0.22569735561737775, "trace": "" }, "children": [ @@ -5221,8 +5221,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 43.323, - "pct_cuda_time": 0.23218329634039186, + "cuda_time_us": 42.102, + "pct_cuda_time": 0.22569735561737775, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -5234,9 +5234,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 36.295, - "cuda_time_us": 9.413, - "pct_cuda_time": 0.05044759985347526, + "cpu_time_us": 37.937, + "cuda_time_us": 10.054, + "pct_cuda_time": 0.05389675581628226, "trace": "" }, "children": [ @@ -5244,8 +5244,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.413, - "pct_cuda_time": 0.05044759985347526, + "cuda_time_us": 10.054, + "pct_cuda_time": 0.05389675581628226, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5255,18 +5255,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 217.144, - "cuda_time_us": 334.366, - "pct_cuda_time": 1.7919857827055252, + "cpu_time_us": 228.862, + "cuda_time_us": 337.18999999999994, + "pct_cuda_time": 1.807583757080984, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 83.394, - "cuda_time_us": 201.941, - "pct_cuda_time": 1.0822733200903694, + "cpu_time_us": 83.585, + "cuda_time_us": 203.033, + "pct_cuda_time": 1.0884046174305984, "trace": "" }, "children": [ @@ -5274,8 +5274,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 201.941, - "pct_cuda_time": 1.0822733200903694, + "cuda_time_us": 203.033, + "pct_cuda_time": 1.0884046174305984, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -5285,9 +5285,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.279, - "cuda_time_us": 13.179, - "pct_cuda_time": 0.07063092727812073, + "cpu_time_us": 27.421, + "cuda_time_us": 13.539, + "pct_cuda_time": 0.0725788916845679, "trace": "" }, "children": [ @@ -5295,8 +5295,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.179, - "pct_cuda_time": 0.07063092727812073, + "cuda_time_us": 13.539, + "pct_cuda_time": 0.0725788916845679, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -5306,9 +5306,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 89.424, - "cuda_time_us": 119.246, - "pct_cuda_time": 0.6390815353370349, + "cpu_time_us": 99.559, + "cuda_time_us": 120.618, + "pct_cuda_time": 0.6466002479658179, "trace": "" }, "children": [ @@ -5316,8 +5316,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 112.797, - "pct_cuda_time": 0.6045190609446986, + "cuda_time_us": 114.05, + "pct_cuda_time": 0.6113909887454736, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5326,8 +5326,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.449, - "pct_cuda_time": 0.03456247439233634, + "cuda_time_us": 6.568, + "pct_cuda_time": 0.035209259220344326, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5341,18 +5341,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1030.018, - "cuda_time_us": 567.894, - "pct_cuda_time": 3.0435450197800358, + "cpu_time_us": 1163.119, + "cuda_time_us": 567.654, + "pct_cuda_time": 3.0430384947419826, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.07, - "cuda_time_us": 9.773, - "pct_cuda_time": 0.05237696731839091, + "cpu_time_us": 34.933, + "cuda_time_us": 9.414, + "pct_cuda_time": 0.05046589011880656, "trace": "" }, "children": [ @@ -5360,8 +5360,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.773, - "pct_cuda_time": 0.05237696731839091, + "cuda_time_us": 9.414, + "pct_cuda_time": 0.05046589011880656, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5371,18 +5371,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 724.872, - "cuda_time_us": 217.718, - "pct_cuda_time": 1.166827849240298, + "cpu_time_us": 832.975, + "cuda_time_us": 221.913, + "pct_cuda_time": 1.1896151555061316, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.924, - "cuda_time_us": 54.661, - "pct_cuda_time": 0.2929476527770967, + "cpu_time_us": 83.505, + "cuda_time_us": 53.989, + "pct_cuda_time": 0.28942032522033645, "trace": "" }, "children": [ @@ -5390,8 +5390,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.661, - "pct_cuda_time": 0.2929476527770967, + "cuda_time_us": 53.989, + "pct_cuda_time": 0.28942032522033645, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -5401,9 +5401,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.059, - "cuda_time_us": 27.236, - "pct_cuda_time": 0.14596736742900798, + "cpu_time_us": 78.618, + "cuda_time_us": 25.552999999999997, + "pct_cuda_time": 0.13698267369936945, "trace": "" }, "children": [ @@ -5411,8 +5411,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -5421,8 +5421,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.406, + "pct_cuda_time": 0.02361936603605924, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -5431,8 +5431,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.89, - "pct_cuda_time": 0.04228530360606818, + "cuda_time_us": 6.528, + "pct_cuda_time": 0.0349948301142521, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -5441,8 +5441,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.448, - "pct_cuda_time": 0.02919776096905697, + "cuda_time_us": 6.008, + "pct_cuda_time": 0.032207251735053094, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -5451,8 +5451,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.045, + "pct_cuda_time": 0.021684143353576858, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -5462,9 +5462,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 428.44, - "cuda_time_us": 93.49799999999999, - "pct_cuda_time": 0.5010888867630117, + "cpu_time_us": 445.887, + "cuda_time_us": 98.626, + "pct_cuda_time": 0.5287071254363094, "trace": "" }, "children": [ @@ -5472,8 +5472,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.407, - "pct_cuda_time": 0.02897802745221935, + "cuda_time_us": 6.649, + "pct_cuda_time": 0.0356434781601811, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -5482,8 +5482,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 88.091, - "pct_cuda_time": 0.4721108593107924, + "cuda_time_us": 91.977, + "pct_cuda_time": 0.4930636472761283, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -5493,9 +5493,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 98.016, - "cuda_time_us": 42.323, - "pct_cuda_time": 0.2268239422711817, + "cpu_time_us": 101.742, + "cuda_time_us": 43.745, + "pct_cuda_time": 0.23450503115011612, "trace": "" }, "children": [ @@ -5503,8 +5503,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.323, - "pct_cuda_time": 0.2268239422711817, + "cuda_time_us": 43.745, + "pct_cuda_time": 0.23450503115011612, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -5516,9 +5516,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 36.745, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.05151947066731728, + "cpu_time_us": 39.92, + "cuda_time_us": 9.853, + "pct_cuda_time": 0.0528192495581688, "trace": "" }, "children": [ @@ -5526,8 +5526,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.05151947066731728, + "cuda_time_us": 9.853, + "pct_cuda_time": 0.0528192495581688, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5537,18 +5537,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 214.44, - "cuda_time_us": 330.79, - "pct_cuda_time": 1.77282073255403, + "cpu_time_us": 233.038, + "cuda_time_us": 326.474, + "pct_cuda_time": 1.7501381995588756, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 84.236, - "cuda_time_us": 196.953, - "pct_cuda_time": 1.0555408619931492, + "cpu_time_us": 95.573, + "cuda_time_us": 193.489, + "pct_cuda_time": 1.037241832716992, "trace": "" }, "children": [ @@ -5556,8 +5556,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 196.953, - "pct_cuda_time": 1.0555408619931492, + "cuda_time_us": 193.489, + "pct_cuda_time": 1.037241832716992, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -5567,9 +5567,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.57, - "cuda_time_us": 13.619, - "pct_cuda_time": 0.0729890430685732, + "cpu_time_us": 26.77, + "cuda_time_us": 13.459, + "pct_cuda_time": 0.07215003347238343, "trace": "" }, "children": [ @@ -5577,8 +5577,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.619, - "pct_cuda_time": 0.0729890430685732, + "cuda_time_us": 13.459, + "pct_cuda_time": 0.07215003347238343, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -5588,9 +5588,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 86.399, - "cuda_time_us": 120.218, - "pct_cuda_time": 0.6442908274923074, + "cpu_time_us": 91.147, + "cuda_time_us": 119.526, + "pct_cuda_time": 0.6407463333695, "trace": "" }, "children": [ @@ -5598,8 +5598,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 113.649, - "pct_cuda_time": 0.6090852306116658, + "cuda_time_us": 112.958, + "pct_cuda_time": 0.6055370741491557, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5608,8 +5608,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.569, - "pct_cuda_time": 0.03520559688064156, + "cuda_time_us": 6.568, + "pct_cuda_time": 0.035209259220344326, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5623,18 +5623,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1051.77, - "cuda_time_us": 569.144, - "pct_cuda_time": 3.050244212366549, + "cpu_time_us": 1098.622, + "cuda_time_us": 568.324, + "pct_cuda_time": 3.046630182269027, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.211, - "cuda_time_us": 9.934, - "pct_cuda_time": 0.053239823323533754, + "cpu_time_us": 42.414, + "cuda_time_us": 9.654, + "pct_cuda_time": 0.05175246475535995, "trace": "" }, "children": [ @@ -5642,8 +5642,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.934, - "pct_cuda_time": 0.053239823323533754, + "cuda_time_us": 9.654, + "pct_cuda_time": 0.05175246475535995, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5653,18 +5653,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 732.494, - "cuda_time_us": 222.833, - "pct_cuda_time": 1.1942409453043081, + "cpu_time_us": 756.071, + "cuda_time_us": 219.29699999999997, + "pct_cuda_time": 1.1755914919676993, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.073, - "cuda_time_us": 53.919, - "pct_cuda_time": 0.2889710120577427, + "cpu_time_us": 89.504, + "cuda_time_us": 54.23, + "pct_cuda_time": 0.2907122605845422, "trace": "" }, "children": [ @@ -5672,8 +5672,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 53.919, - "pct_cuda_time": 0.2889710120577427, + "cuda_time_us": 54.23, + "pct_cuda_time": 0.2907122605845422, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -5683,9 +5683,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.169, - "cuda_time_us": 26.715000000000003, - "pct_cuda_time": 0.1431751439589495, + "cpu_time_us": 74.771, + "cuda_time_us": 26.715, + "pct_cuda_time": 0.14321183923134878, "trace": "" }, "children": [ @@ -5693,8 +5693,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -5703,8 +5703,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.485, + "pct_cuda_time": 0.024042863520591404, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -5713,8 +5713,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.129, - "pct_cuda_time": 0.038206835159399245, + "cuda_time_us": 7.731, + "pct_cuda_time": 0.04144378547997595, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -5723,8 +5723,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.488, - "pct_cuda_time": 0.029412135131825373, + "cuda_time_us": 5.247, + "pct_cuda_time": 0.028127737991648402, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -5733,8 +5733,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.926, - "pct_cuda_time": 0.026400178144929263, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -5744,9 +5744,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 440.788, - "cuda_time_us": 94.05799999999999, - "pct_cuda_time": 0.5040901250417694, + "cpu_time_us": 452.316, + "cuda_time_us": 95.82, + "pct_cuda_time": 0.5136649236439393, "trace": "" }, "children": [ @@ -5754,8 +5754,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.487, - "pct_cuda_time": 0.029406775777756165, + "cuda_time_us": 5.687, + "pct_cuda_time": 0.030486458158662945, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -5764,8 +5764,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 88.571, - "pct_cuda_time": 0.47468334926401323, + "cuda_time_us": 90.133, + "pct_cuda_time": 0.48317846548527643, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -5775,9 +5775,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 101.0, - "cuda_time_us": 48.141, - "pct_cuda_time": 0.25800466424584645, + "cpu_time_us": 103.845, + "cuda_time_us": 42.532, + "pct_cuda_time": 0.2280024685078692, "trace": "" }, "children": [ @@ -5785,8 +5785,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 48.141, - "pct_cuda_time": 0.25800466424584645, + "cuda_time_us": 42.532, + "pct_cuda_time": 0.2280024685078692, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -5798,9 +5798,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.698, - "cuda_time_us": 10.134, - "pct_cuda_time": 0.054311694137375785, + "cpu_time_us": 38.728, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "" }, "children": [ @@ -5808,8 +5808,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.134, - "pct_cuda_time": 0.054311694137375785, + "cuda_time_us": 9.413, + "pct_cuda_time": 0.050460529391154266, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5819,18 +5819,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 224.606, - "cuda_time_us": 326.243, - "pct_cuda_time": 1.7484517496013312, + "cpu_time_us": 237.696, + "cuda_time_us": 329.96, + "pct_cuda_time": 1.7688256961548137, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 86.009, - "cuda_time_us": 194.64, - "pct_cuda_time": 1.0431446760310659, + "cpu_time_us": 99.639, + "cuda_time_us": 196.083, + "pct_cuda_time": 1.0511475602470732, "trace": "" }, "children": [ @@ -5838,8 +5838,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 194.64, - "pct_cuda_time": 1.0431446760310659, + "cuda_time_us": 196.083, + "pct_cuda_time": 1.0511475602470732, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -5849,9 +5849,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.89, - "cuda_time_us": 13.219, - "pct_cuda_time": 0.07084530144088913, + "cpu_time_us": 27.15, + "cuda_time_us": 14.22, + "pct_cuda_time": 0.07622954721578812, "trace": "" }, "children": [ @@ -5859,8 +5859,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.219, - "pct_cuda_time": 0.07084530144088913, + "cuda_time_us": 14.22, + "pct_cuda_time": 0.07622954721578812, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -5870,9 +5870,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 93.25, - "cuda_time_us": 118.384, - "pct_cuda_time": 0.6344617721293758, + "cpu_time_us": 90.185, + "cuda_time_us": 119.657, + "pct_cuda_time": 0.6414485886919521, "trace": "" }, "children": [ @@ -5880,8 +5880,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 112.016, - "pct_cuda_time": 0.6003334054166456, + "cuda_time_us": 113.128, + "pct_cuda_time": 0.6064483978500478, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5890,8 +5890,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.368, - "pct_cuda_time": 0.03412836671273032, + "cuda_time_us": 6.529, + "pct_cuda_time": 0.03500019084190441, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -5905,18 +5905,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1039.582, - "cuda_time_us": 566.03, - "pct_cuda_time": 3.033555183795028, + "cpu_time_us": 1102.818, + "cuda_time_us": 573.953, + "pct_cuda_time": 3.0768057182238566, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.512, - "cuda_time_us": 9.654, - "pct_cuda_time": 0.05173920418415491, + "cpu_time_us": 34.151, + "cuda_time_us": 9.894, + "pct_cuda_time": 0.05303903939191334, "trace": "" }, "children": [ @@ -5924,8 +5924,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.654, - "pct_cuda_time": 0.05173920418415491, + "cuda_time_us": 9.894, + "pct_cuda_time": 0.05303903939191334, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -5935,18 +5935,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 719.764, - "cuda_time_us": 220.63000000000002, - "pct_cuda_time": 1.1824342882898382, + "cpu_time_us": 773.847, + "cuda_time_us": 225.31799999999998, + "pct_cuda_time": 1.2078684331622327, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 83.975, - "cuda_time_us": 54.45, - "pct_cuda_time": 0.2918168290684934, + "cpu_time_us": 98.837, + "cuda_time_us": 54.75, + "pct_cuda_time": 0.2934998389637412, "trace": "" }, "children": [ @@ -5954,8 +5954,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.45, - "pct_cuda_time": 0.2918168290684934, + "cuda_time_us": 54.75, + "pct_cuda_time": 0.2934998389637412, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -5965,9 +5965,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 74.26, - "cuda_time_us": 26.994999999999997, - "pct_cuda_time": 0.14467576309832833, + "cpu_time_us": 76.694, + "cuda_time_us": 26.073999999999998, + "pct_cuda_time": 0.13977561280622078, "trace": "" }, "children": [ @@ -5975,8 +5975,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.0246915115665204, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -5985,8 +5985,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.486, + "pct_cuda_time": 0.024048224248243707, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -5996,7 +5996,7 @@ "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, "cuda_time_us": 7.21, - "pct_cuda_time": 0.03864094283900527, + "pct_cuda_time": 0.038650846373124645, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -6005,8 +6005,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.807, - "pct_cuda_time": 0.031121769079903414, + "cuda_time_us": 5.727, + "pct_cuda_time": 0.030700887264755176, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -6015,8 +6015,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.045, + "pct_cuda_time": 0.021684143353576858, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -6026,9 +6026,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 424.514, - "cuda_time_us": 95.10000000000001, - "pct_cuda_time": 0.5096745719818865, + "cpu_time_us": 457.514, + "cuda_time_us": 95.02099999999999, + "pct_cuda_time": 0.509381702249747, "trace": "" }, "children": [ @@ -6036,8 +6036,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.528, - "pct_cuda_time": 0.029626509294593775, + "cuda_time_us": 6.329, + "pct_cuda_time": 0.03392804531144325, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6046,8 +6046,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 89.572, - "pct_cuda_time": 0.4800480626872927, + "cuda_time_us": 88.692, + "pct_cuda_time": 0.47545365693830377, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6057,9 +6057,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.661, - "cuda_time_us": 44.085, - "pct_cuda_time": 0.23626712414113, + "cpu_time_us": 101.451, + "cuda_time_us": 49.473, + "pct_cuda_time": 0.2652112791425236, "trace": "" }, "children": [ @@ -6067,8 +6067,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 44.085, - "pct_cuda_time": 0.23626712414113, + "cuda_time_us": 49.473, + "pct_cuda_time": 0.2652112791425236, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -6080,9 +6080,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.727, - "cuda_time_us": 9.694, - "pct_cuda_time": 0.05195357834692332, + "cpu_time_us": 38.267, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "" }, "children": [ @@ -6090,8 +6090,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.694, - "pct_cuda_time": 0.05195357834692332, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6101,18 +6101,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 225.767, - "cuda_time_us": 326.052, - "pct_cuda_time": 1.7474281129741118, + "cpu_time_us": 233.69, + "cuda_time_us": 328.807, + "pct_cuda_time": 1.762644777171705, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.267, - "cuda_time_us": 196.272, - "pct_cuda_time": 1.051891141872017, + "cpu_time_us": 89.764, + "cuda_time_us": 197.955, + "pct_cuda_time": 1.0611828424121899, "trace": "" }, "children": [ @@ -6120,8 +6120,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 196.272, - "pct_cuda_time": 1.051891141872017, + "cuda_time_us": 197.955, + "pct_cuda_time": 1.0611828424121899, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -6131,9 +6131,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.251, - "cuda_time_us": 12.978, - "pct_cuda_time": 0.06955369711020949, + "cpu_time_us": 26.269, + "cuda_time_us": 13.218, + "pct_cuda_time": 0.07085809810817774, "trace": "" }, "children": [ @@ -6141,8 +6141,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 12.978, - "pct_cuda_time": 0.06955369711020949, + "cuda_time_us": 13.218, + "pct_cuda_time": 0.07085809810817774, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -6152,9 +6152,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 94.101, - "cuda_time_us": 116.80199999999999, - "pct_cuda_time": 0.6259832739918854, + "cpu_time_us": 98.558, + "cuda_time_us": 117.634, + "pct_cuda_time": 0.6306038366513376, "trace": "" }, "children": [ @@ -6162,8 +6162,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 110.234, - "pct_cuda_time": 0.590783036465313, + "cuda_time_us": 111.306, + "pct_cuda_time": 0.5966811520675466, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -6172,8 +6172,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.568, - "pct_cuda_time": 0.035200237526572344, + "cuda_time_us": 6.328, + "pct_cuda_time": 0.03392268458379095, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -6187,18 +6187,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1040.914, - "cuda_time_us": 559.521, - "pct_cuda_time": 2.998671148158539, + "cpu_time_us": 1064.221, + "cuda_time_us": 560.814, + "pct_cuda_time": 3.006371117600211, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.002, - "cuda_time_us": 9.934, - "pct_cuda_time": 0.053239823323533754, + "cpu_time_us": 35.042, + "cuda_time_us": 9.654, + "pct_cuda_time": 0.05175246475535995, "trace": "" }, "children": [ @@ -6206,8 +6206,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.934, - "pct_cuda_time": 0.053239823323533754, + "cuda_time_us": 9.654, + "pct_cuda_time": 0.05175246475535995, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6217,18 +6217,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 726.825, - "cuda_time_us": 221.98299999999998, - "pct_cuda_time": 1.1896854943454793, + "cpu_time_us": 756.03, + "cuda_time_us": 221.953, + "pct_cuda_time": 1.1898295846122238, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.312, - "cuda_time_us": 54.18, - "pct_cuda_time": 0.2903698034698066, + "cpu_time_us": 83.264, + "cuda_time_us": 54.27, + "pct_cuda_time": 0.29092668969063445, "trace": "" }, "children": [ @@ -6236,8 +6236,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.18, - "pct_cuda_time": 0.2903698034698066, + "cuda_time_us": 54.27, + "pct_cuda_time": 0.29092668969063445, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -6247,9 +6247,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.28, - "cuda_time_us": 27.876999999999995, - "pct_cuda_time": 0.14940271338737168, + "cpu_time_us": 77.716, + "cuda_time_us": 28.076999999999998, + "pct_cuda_time": 0.15051315029378923, "trace": "" }, "children": [ @@ -6257,8 +6257,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.0246915115665204, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -6267,8 +6267,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.486, - "pct_cuda_time": 0.024042062354476788, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -6277,8 +6277,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.851, - "pct_cuda_time": 0.04207628879736899, + "cuda_time_us": 8.572, + "pct_cuda_time": 0.0459521574355651, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -6287,8 +6287,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.767, - "pct_cuda_time": 0.030907394917135008, + "cuda_time_us": 5.648, + "pct_cuda_time": 0.030277389780223016, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -6297,8 +6297,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.007, - "pct_cuda_time": 0.02683428582453528, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -6308,9 +6308,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 435.66, - "cuda_time_us": 95.901, - "pct_cuda_time": 0.5139674145913238, + "cpu_time_us": 450.313, + "cuda_time_us": 95.06, + "pct_cuda_time": 0.509590770628187, "trace": "" }, "children": [ @@ -6318,8 +6318,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.687, - "pct_cuda_time": 0.030478646591598192, + "cuda_time_us": 5.647, + "pct_cuda_time": 0.030272029052570714, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6328,8 +6328,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 90.214, - "pct_cuda_time": 0.4834887679997256, + "cuda_time_us": 89.413, + "pct_cuda_time": 0.47931874157561627, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6339,9 +6339,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.671, - "cuda_time_us": 44.025, - "pct_cuda_time": 0.2359455628969774, + "cpu_time_us": 105.458, + "cuda_time_us": 44.546, + "pct_cuda_time": 0.23879897399961303, "trace": "" }, "children": [ @@ -6349,8 +6349,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 44.025, - "pct_cuda_time": 0.2359455628969774, + "cuda_time_us": 44.546, + "pct_cuda_time": 0.23879897399961303, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -6362,9 +6362,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.227, - "cuda_time_us": 9.734, - "pct_cuda_time": 0.052167952509691716, + "cpu_time_us": 36.915, + "cuda_time_us": 10.414, + "pct_cuda_time": 0.05582661777111234, "trace": "" }, "children": [ @@ -6372,8 +6372,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.734, - "pct_cuda_time": 0.052167952509691716, + "cuda_time_us": 10.414, + "pct_cuda_time": 0.05582661777111234, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6383,18 +6383,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 219.748, - "cuda_time_us": 317.87, - "pct_cuda_time": 1.7035778779798345, + "cpu_time_us": 215.011, + "cuda_time_us": 318.793, + "pct_cuda_time": 1.7089624504615148, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 83.165, - "cuda_time_us": 187.478, - "pct_cuda_time": 1.0047609821873829, + "cpu_time_us": 83.334, + "cuda_time_us": 186.218, + "pct_cuda_time": 0.9982639819570769, "trace": "" }, "children": [ @@ -6402,8 +6402,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 187.478, - "pct_cuda_time": 1.0047609821873829, + "cuda_time_us": 186.218, + "pct_cuda_time": 0.9982639819570769, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -6413,9 +6413,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.85, - "cuda_time_us": 13.179, - "pct_cuda_time": 0.07063092727812073, + "cpu_time_us": 26.299, + "cuda_time_us": 13.258, + "pct_cuda_time": 0.07107252721426996, "trace": "" }, "children": [ @@ -6423,8 +6423,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.179, - "pct_cuda_time": 0.07063092727812073, + "cuda_time_us": 13.258, + "pct_cuda_time": 0.07107252721426996, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -6434,9 +6434,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 91.476, - "cuda_time_us": 117.21300000000001, - "pct_cuda_time": 0.6281859685143308, + "cpu_time_us": 86.759, + "cuda_time_us": 119.317, + "pct_cuda_time": 0.6396259412901681, "trace": "" }, "children": [ @@ -6444,8 +6444,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 110.724, - "pct_cuda_time": 0.593409119959226, + "cuda_time_us": 112.467, + "pct_cuda_time": 0.6029049568718736, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -6454,8 +6454,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.489, - "pct_cuda_time": 0.034776848555104746, + "cuda_time_us": 6.85, + "pct_cuda_time": 0.03672098441829456, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -6469,18 +6469,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1038.139, - "cuda_time_us": 566.0409999999999, - "pct_cuda_time": 3.0336141366897893, + "cpu_time_us": 1099.754, + "cuda_time_us": 560.333, + "pct_cuda_time": 3.003792607599452, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.121, - "cuda_time_us": 9.734, - "pct_cuda_time": 0.052167952509691716, + "cpu_time_us": 34.442, + "cuda_time_us": 9.573, + "pct_cuda_time": 0.05131824581552319, "trace": "" }, "children": [ @@ -6488,8 +6488,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.734, - "pct_cuda_time": 0.052167952509691716, + "cuda_time_us": 9.573, + "pct_cuda_time": 0.05131824581552319, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6499,18 +6499,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 723.76, - "cuda_time_us": 223.214, - "pct_cuda_time": 1.1962828592046773, + "cpu_time_us": 775.299, + "cuda_time_us": 219.04899999999998, + "pct_cuda_time": 1.1742620315099277, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 84.496, - "cuda_time_us": 54.029, - "pct_cuda_time": 0.2895605410053559, + "cpu_time_us": 87.441, + "cuda_time_us": 53.9, + "pct_cuda_time": 0.2889432204592813, "trace": "" }, "children": [ @@ -6518,8 +6518,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.029, - "pct_cuda_time": 0.2895605410053559, + "cuda_time_us": 53.9, + "pct_cuda_time": 0.2889432204592813, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -6529,9 +6529,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 73.73, - "cuda_time_us": 28.076999999999998, - "pct_cuda_time": 0.1504745842012137, + "cpu_time_us": 75.022, + "cuda_time_us": 26.314, + "pct_cuda_time": 0.14106218744277418, "trace": "" }, "children": [ @@ -6539,8 +6539,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -6549,8 +6549,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.525, + "pct_cuda_time": 0.024257292626683635, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -6559,8 +6559,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 8.372, - "pct_cuda_time": 0.04486851226742748, + "cuda_time_us": 6.729, + "pct_cuda_time": 0.03607233637236556, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -6569,8 +6569,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.767, - "pct_cuda_time": 0.030907394917135008, + "cuda_time_us": 5.728, + "pct_cuda_time": 0.03070624799240748, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -6579,8 +6579,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.02490594067261263, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -6590,9 +6590,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 429.371, - "cuda_time_us": 96.663, - "pct_cuda_time": 0.5180512423920619, + "cpu_time_us": 473.258, + "cuda_time_us": 97.18299999999999, + "pct_cuda_time": 0.5209715954340322, "trace": "" }, "children": [ @@ -6600,8 +6600,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.648, - "pct_cuda_time": 0.030269631782898994, + "cuda_time_us": 5.567, + "pct_cuda_time": 0.029843170840386252, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6610,8 +6610,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 91.015, - "pct_cuda_time": 0.487781610609163, + "cuda_time_us": 91.616, + "pct_cuda_time": 0.49112842459364586, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6621,9 +6621,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.339, - "cuda_time_us": 44.445, - "pct_cuda_time": 0.23819649160604564, + "cpu_time_us": 100.28, + "cuda_time_us": 41.652, + "pct_cuda_time": 0.22328502817384016, "trace": "" }, "children": [ @@ -6631,8 +6631,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 44.445, - "pct_cuda_time": 0.23819649160604564, + "cuda_time_us": 41.652, + "pct_cuda_time": 0.22328502817384016, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -6644,9 +6644,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.597, - "cuda_time_us": 9.814, - "pct_cuda_time": 0.052596700835228535, + "cpu_time_us": 38.598, + "cuda_time_us": 10.254, + "pct_cuda_time": 0.05496890134674342, "trace": "" }, "children": [ @@ -6654,8 +6654,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.814, - "pct_cuda_time": 0.052596700835228535, + "cuda_time_us": 10.254, + "pct_cuda_time": 0.05496890134674342, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6665,18 +6665,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 219.338, - "cuda_time_us": 323.279, - "pct_cuda_time": 1.732566624140192, + "cpu_time_us": 228.602, + "cuda_time_us": 321.457, + "pct_cuda_time": 1.7232434289272576, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.347, - "cuda_time_us": 192.917, - "pct_cuda_time": 1.0339105089698168, + "cpu_time_us": 93.72, + "cuda_time_us": 191.345, + "pct_cuda_time": 1.0257484326304487, "trace": "" }, "children": [ @@ -6684,8 +6684,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 192.917, - "pct_cuda_time": 1.0339105089698168, + "cuda_time_us": 191.345, + "pct_cuda_time": 1.0257484326304487, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -6695,9 +6695,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.78, - "cuda_time_us": 13.579, - "pct_cuda_time": 0.07277466890580479, + "cpu_time_us": 27.621, + "cuda_time_us": 13.459, + "pct_cuda_time": 0.07215003347238343, "trace": "" }, "children": [ @@ -6705,8 +6705,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.579, - "pct_cuda_time": 0.07277466890580479, + "cuda_time_us": 13.459, + "pct_cuda_time": 0.07215003347238343, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -6716,9 +6716,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 89.073, - "cuda_time_us": 116.78299999999999, - "pct_cuda_time": 0.6258814462645703, + "cpu_time_us": 89.134, + "cuda_time_us": 116.65299999999999, + "pct_cuda_time": 0.6253449628244255, "trace": "" }, "children": [ @@ -6726,8 +6726,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 110.094, - "pct_cuda_time": 0.5900327268956236, + "cuda_time_us": 110.204, + "pct_cuda_time": 0.5907736301947056, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -6736,8 +6736,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.689, - "pct_cuda_time": 0.03584871936894678, + "cuda_time_us": 6.449, + "pct_cuda_time": 0.03457133262971994, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -6751,18 +6751,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1059.412, - "cuda_time_us": 564.609, - "pct_cuda_time": 3.0259395416626806, + "cpu_time_us": 1122.428, + "cuda_time_us": 555.145, + "pct_cuda_time": 2.9759811525392896, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.3, - "cuda_time_us": 9.252, - "pct_cuda_time": 0.04958474384833243, + "cpu_time_us": 33.35, + "cuda_time_us": 9.453, + "pct_cuda_time": 0.05067495849724649, "trace": "" }, "children": [ @@ -6770,8 +6770,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.252, - "pct_cuda_time": 0.04958474384833243, + "cuda_time_us": 9.453, + "pct_cuda_time": 0.05067495849724649, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6781,18 +6781,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 740.936, - "cuda_time_us": 222.79500000000002, - "pct_cuda_time": 1.1940372898496783, + "cpu_time_us": 788.679, + "cuda_time_us": 218.989, + "pct_cuda_time": 1.1739403878507895, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 85.658, - "cuda_time_us": 54.441, - "pct_cuda_time": 0.29176859488187046, + "cpu_time_us": 90.765, + "cuda_time_us": 52.187, + "pct_cuda_time": 0.2797602939908815, "trace": "" }, "children": [ @@ -6800,8 +6800,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 54.441, - "pct_cuda_time": 0.29176859488187046, + "cuda_time_us": 52.187, + "pct_cuda_time": 0.2797602939908815, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -6811,9 +6811,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 86.018, - "cuda_time_us": 27.115000000000002, - "pct_cuda_time": 0.14531888558663358, + "cpu_time_us": 76.685, + "cuda_time_us": 27.395999999999997, + "pct_cuda_time": 0.14686249476256902, "trace": "" }, "children": [ @@ -6821,8 +6821,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -6831,8 +6831,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.486, - "pct_cuda_time": 0.024042062354476788, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -6841,8 +6841,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.33, - "pct_cuda_time": 0.03928406532731049, + "cuda_time_us": 7.45, + "pct_cuda_time": 0.03993742100967803, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -6851,8 +6851,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.687, - "pct_cuda_time": 0.030478646591598192, + "cuda_time_us": 6.008, + "pct_cuda_time": 0.032207251735053094, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -6861,8 +6861,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -6872,9 +6872,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 431.955, - "cuda_time_us": 96.743, - "pct_cuda_time": 0.5184799907175988, + "cpu_time_us": 483.482, + "cuda_time_us": 97.664, + "pct_cuda_time": 0.5235501054347913, "trace": "" }, "children": [ @@ -6882,8 +6882,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.568, - "pct_cuda_time": 0.02984088345736218, + "cuda_time_us": 5.607, + "pct_cuda_time": 0.030057599946478483, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6892,8 +6892,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 91.175, - "pct_cuda_time": 0.4886391072602365, + "cuda_time_us": 92.057, + "pct_cuda_time": 0.49349250548831275, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -6903,9 +6903,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.99, - "cuda_time_us": 44.496, - "pct_cuda_time": 0.2384698186635754, + "cpu_time_us": 102.153, + "cuda_time_us": 41.742, + "pct_cuda_time": 0.22376749366254767, "trace": "" }, "children": [ @@ -6913,8 +6913,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 44.496, - "pct_cuda_time": 0.2384698186635754, + "cuda_time_us": 41.742, + "pct_cuda_time": 0.22376749366254767, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -6926,9 +6926,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 40.741, - "cuda_time_us": 10.495, - "pct_cuda_time": 0.056246420956360656, + "cpu_time_us": 41.693, + "cuda_time_us": 10.615, + "pct_cuda_time": 0.0569041240292258, "trace": "" }, "children": [ @@ -6936,8 +6936,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.495, - "pct_cuda_time": 0.056246420956360656, + "cuda_time_us": 10.615, + "pct_cuda_time": 0.0569041240292258, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -6947,18 +6947,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 223.534, - "cuda_time_us": 322.067, - "pct_cuda_time": 1.7260710870083094, + "cpu_time_us": 235.282, + "cuda_time_us": 316.088, + "pct_cuda_time": 1.6944616821620282, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 87.0, - "cuda_time_us": 193.327, - "pct_cuda_time": 1.036107844138193, + "cpu_time_us": 85.388, + "cuda_time_us": 188.12, + "pct_cuda_time": 1.0084600859517625, "trace": "" }, "children": [ @@ -6966,8 +6966,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 193.327, - "pct_cuda_time": 1.036107844138193, + "cuda_time_us": 188.12, + "pct_cuda_time": 1.0084600859517625, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -6977,9 +6977,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.801, + "cpu_time_us": 37.055, "cuda_time_us": 13.259, - "pct_cuda_time": 0.07105967560365754, + "pct_cuda_time": 0.07107788794192228, "trace": "" }, "children": [ @@ -6988,7 +6988,7 @@ "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, "cuda_time_us": 13.259, - "pct_cuda_time": 0.07105967560365754, + "pct_cuda_time": 0.07107788794192228, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -6998,9 +6998,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.275, - "cuda_time_us": 115.481, - "pct_cuda_time": 0.6189035672664587, + "cpu_time_us": 93.6, + "cuda_time_us": 114.709, + "pct_cuda_time": 0.6149237082683432, "trace": "" }, "children": [ @@ -7008,8 +7008,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 109.393, - "pct_cuda_time": 0.5862758196931073, + "cuda_time_us": 108.221, + "pct_cuda_time": 0.5801433072601834, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7018,8 +7018,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.088, - "pct_cuda_time": 0.03262774757335147, + "cuda_time_us": 6.488, + "pct_cuda_time": 0.03478040100815987, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7033,18 +7033,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1053.263, - "cuda_time_us": 555.864, - "pct_cuda_time": 2.979071990327438, + "cpu_time_us": 1156.308, + "cuda_time_us": 551.741, + "pct_cuda_time": 2.957733235610841, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 33.35, - "cuda_time_us": 9.853, - "pct_cuda_time": 0.05280571564392773, + "cpu_time_us": 38.828, + "cuda_time_us": 9.694, + "pct_cuda_time": 0.05196689386145219, "trace": "" }, "children": [ @@ -7052,8 +7052,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.853, - "pct_cuda_time": 0.05280571564392773, + "cuda_time_us": 9.694, + "pct_cuda_time": 0.05196689386145219, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7063,18 +7063,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 737.011, - "cuda_time_us": 220.25000000000003, - "pct_cuda_time": 1.1803977337435383, + "cpu_time_us": 824.463, + "cuda_time_us": 216.42499999999998, + "pct_cuda_time": 1.1601954821502773, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 85.387, - "cuda_time_us": 51.727, - "pct_cuda_time": 0.27722330793803407, + "cpu_time_us": 88.602, + "cuda_time_us": 51.265, + "pct_cuda_time": 0.2748177030954556, "trace": "" }, "children": [ @@ -7082,8 +7082,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 51.727, - "pct_cuda_time": 0.27722330793803407, + "cuda_time_us": 51.265, + "pct_cuda_time": 0.2748177030954556, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -7093,9 +7093,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 76.344, - "cuda_time_us": 26.955000000000002, - "pct_cuda_time": 0.14446138893555993, + "cpu_time_us": 77.065, + "cuda_time_us": 26.354999999999997, + "pct_cuda_time": 0.1412819772765187, "trace": "" }, "children": [ @@ -7104,7 +7104,7 @@ "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -7113,8 +7113,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.565, - "pct_cuda_time": 0.024465451325944392, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -7123,8 +7123,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.05, - "pct_cuda_time": 0.03778344618793164, + "cuda_time_us": 7.37, + "pct_cuda_time": 0.03950856279749357, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -7133,8 +7133,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.928, - "pct_cuda_time": 0.031770250922277844, + "cuda_time_us": 5.568, + "pct_cuda_time": 0.029848531568038555, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -7143,8 +7143,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.125, + "pct_cuda_time": 0.022113001565761323, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -7154,9 +7154,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 440.117, - "cuda_time_us": 97.66300000000001, - "pct_cuda_time": 0.5234105964612722, + "cpu_time_us": 521.229, + "cuda_time_us": 99.628, + "pct_cuda_time": 0.5340785745439198, "trace": "" }, "children": [ @@ -7164,8 +7164,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.727, - "pct_cuda_time": 0.030693020754366602, + "cuda_time_us": 6.97, + "pct_cuda_time": 0.03736427173657125, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -7174,8 +7174,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 91.936, - "pct_cuda_time": 0.49271757570690555, + "cuda_time_us": 92.658, + "pct_cuda_time": 0.4967143028073486, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -7185,9 +7185,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 99.148, - "cuda_time_us": 43.905, - "pct_cuda_time": 0.2353024404086722, + "cpu_time_us": 100.83, + "cuda_time_us": 39.177, + "pct_cuda_time": 0.21001722723438337, "trace": "" }, "children": [ @@ -7195,8 +7195,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 43.905, - "pct_cuda_time": 0.2353024404086722, + "cuda_time_us": 39.177, + "pct_cuda_time": 0.21001722723438337, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -7208,9 +7208,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.286, - "cuda_time_us": 9.894, - "pct_cuda_time": 0.053025449160765355, + "cpu_time_us": 37.906, + "cuda_time_us": 10.735, + "pct_cuda_time": 0.05754741134750249, "trace": "" }, "children": [ @@ -7218,8 +7218,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.894, - "pct_cuda_time": 0.053025449160765355, + "cuda_time_us": 10.735, + "pct_cuda_time": 0.05754741134750249, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7229,18 +7229,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 224.315, - "cuda_time_us": 315.867, - "pct_cuda_time": 1.6928430917792066, + "cpu_time_us": 230.986, + "cuda_time_us": 314.887, + "pct_cuda_time": 1.6880234482516088, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.487, - "cuda_time_us": 186.217, - "pct_cuda_time": 0.9980028367061088, + "cpu_time_us": 84.777, + "cuda_time_us": 185.336, + "pct_cuda_time": 0.9935358201677432, "trace": "" }, "children": [ @@ -7248,8 +7248,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 186.217, - "pct_cuda_time": 0.9980028367061088, + "cuda_time_us": 185.336, + "pct_cuda_time": 0.9935358201677432, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -7259,9 +7259,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.512, - "cuda_time_us": 13.218, - "pct_cuda_time": 0.07083994208681993, + "cpu_time_us": 27.041, + "cuda_time_us": 13.219, + "pct_cuda_time": 0.07086345883583005, "trace": "" }, "children": [ @@ -7269,8 +7269,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.218, - "pct_cuda_time": 0.07083994208681993, + "cuda_time_us": 13.219, + "pct_cuda_time": 0.07086345883583005, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -7280,9 +7280,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 92.428, - "cuda_time_us": 116.432, - "pct_cuda_time": 0.6240003129862777, + "cpu_time_us": 99.94, + "cuda_time_us": 116.332, + "pct_cuda_time": 0.6236241692480354, "trace": "" }, "children": [ @@ -7290,8 +7290,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 109.343, - "pct_cuda_time": 0.5860078519896468, + "cuda_time_us": 109.883, + "pct_cuda_time": 0.5890528366183155, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7300,8 +7300,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 7.089, - "pct_cuda_time": 0.037992460996630846, + "cuda_time_us": 6.449, + "pct_cuda_time": 0.03457133262971994, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7315,18 +7315,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1073.152, - "cuda_time_us": 565.168, - "pct_cuda_time": 3.028935420587369, + "cpu_time_us": 1083.67, + "cuda_time_us": 558.5509999999999, + "pct_cuda_time": 2.9942397909230425, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.363, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cpu_time_us": 35.082, + "cuda_time_us": 10.054, + "pct_cuda_time": 0.05389675581628226, "trace": "" }, "children": [ @@ -7334,8 +7334,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cuda_time_us": 10.054, + "pct_cuda_time": 0.05389675581628226, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7345,18 +7345,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 751.081, - "cuda_time_us": 223.914, - "pct_cuda_time": 1.2000344070531241, + "cpu_time_us": 755.339, + "cuda_time_us": 219.46999999999997, + "pct_cuda_time": 1.1765188978515484, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 89.103, - "cuda_time_us": 53.509, - "pct_cuda_time": 0.28677367688936656, + "cpu_time_us": 85.367, + "cuda_time_us": 51.146, + "pct_cuda_time": 0.27417977650483116, "trace": "" }, "children": [ @@ -7364,8 +7364,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 53.509, - "pct_cuda_time": 0.28677367688936656, + "cuda_time_us": 51.146, + "pct_cuda_time": 0.27417977650483116, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -7375,9 +7375,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 74.892, - "cuda_time_us": 27.876, - "pct_cuda_time": 0.14939735403330248, + "cpu_time_us": 74.832, + "cuda_time_us": 26.714999999999996, + "pct_cuda_time": 0.14321183923134878, "trace": "" }, "children": [ @@ -7385,8 +7385,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "cuda_time_us": 4.846, + "pct_cuda_time": 0.02597808620307379, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -7395,8 +7395,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.024256436517245194, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -7405,8 +7405,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.41, - "pct_cuda_time": 0.039712813652847304, + "cuda_time_us": 6.889, + "pct_cuda_time": 0.03693005279673448, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -7415,8 +7415,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 6.248, - "pct_cuda_time": 0.033485244224425094, + "cuda_time_us": 5.728, + "pct_cuda_time": 0.03070624799240748, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -7425,8 +7425,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.966, - "pct_cuda_time": 0.026614552307697666, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -7436,9 +7436,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 447.869, - "cuda_time_us": 97.18299999999999, - "pct_cuda_time": 0.5208381065080512, + "cpu_time_us": 453.108, + "cuda_time_us": 98.70599999999999, + "pct_cuda_time": 0.5291359836484938, "trace": "" }, "children": [ @@ -7446,8 +7446,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.728, - "pct_cuda_time": 0.030698380108435806, + "cuda_time_us": 5.567, + "pct_cuda_time": 0.029843170840386252, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -7456,8 +7456,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 91.455, - "pct_cuda_time": 0.4901397263996154, + "cuda_time_us": 93.139, + "pct_cuda_time": 0.4992928128081076, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -7467,9 +7467,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 102.123, - "cuda_time_us": 45.346, - "pct_cuda_time": 0.243025269622404, + "cpu_time_us": 102.213, + "cuda_time_us": 42.903, + "pct_cuda_time": 0.22999129846687466, "trace": "" }, "children": [ @@ -7477,8 +7477,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 45.346, - "pct_cuda_time": 0.243025269622404, + "cuda_time_us": 42.903, + "pct_cuda_time": 0.22999129846687466, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -7490,9 +7490,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.388, - "cuda_time_us": 10.775, - "pct_cuda_time": 0.057747040095739506, + "cpu_time_us": 46.57, + "cuda_time_us": 10.455, + "pct_cuda_time": 0.05604640760485688, "trace": "" }, "children": [ @@ -7500,8 +7500,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.775, - "pct_cuda_time": 0.057747040095739506, + "cuda_time_us": 10.455, + "pct_cuda_time": 0.05604640760485688, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7511,18 +7511,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 226.098, - "cuda_time_us": 320.625, - "pct_cuda_time": 1.7183428984405082, + "cpu_time_us": 224.346, + "cuda_time_us": 318.572, + "pct_cuda_time": 1.7077777296503553, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 86.199, - "cuda_time_us": 192.607, - "pct_cuda_time": 1.0322491092083619, + "cpu_time_us": 87.951, + "cuda_time_us": 189.883, + "pct_cuda_time": 1.0179110488027776, "trace": "" }, "children": [ @@ -7530,8 +7530,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 192.607, - "pct_cuda_time": 1.0322491092083619, + "cuda_time_us": 189.883, + "pct_cuda_time": 1.0179110488027776, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -7541,9 +7541,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.57, - "cuda_time_us": 14.06, - "pct_cuda_time": 0.07535251821309488, + "cpu_time_us": 26.73, + "cuda_time_us": 13.338, + "pct_cuda_time": 0.07150138542645443, "trace": "" }, "children": [ @@ -7551,8 +7551,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 14.06, - "pct_cuda_time": 0.07535251821309488, + "cuda_time_us": 13.338, + "pct_cuda_time": 0.07150138542645443, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -7562,9 +7562,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 94.992, - "cuda_time_us": 113.958, - "pct_cuda_time": 0.6107412710190517, + "cpu_time_us": 89.975, + "cuda_time_us": 115.351, + "pct_cuda_time": 0.6183652954211235, "trace": "" }, "children": [ @@ -7572,8 +7572,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 107.469, - "pct_cuda_time": 0.5759644224639469, + "cuda_time_us": 108.261, + "pct_cuda_time": 0.5803577363662755, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7582,8 +7582,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.489, - "pct_cuda_time": 0.034776848555104746, + "cuda_time_us": 7.09, + "pct_cuda_time": 0.03800755905484794, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7597,18 +7597,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1092.201, - "cuda_time_us": 552.309, - "pct_cuda_time": 2.9600194866113956, + "cpu_time_us": 1088.337, + "cuda_time_us": 550.939, + "pct_cuda_time": 2.953433932033691, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.152, - "cuda_time_us": 9.774, - "pct_cuda_time": 0.05238232667246012, + "cpu_time_us": 33.58, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "" }, "children": [ @@ -7616,8 +7616,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.774, - "pct_cuda_time": 0.05238232667246012, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7627,18 +7627,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 749.829, - "cuda_time_us": 218.03599999999997, - "pct_cuda_time": 1.1685321238343067, + "cpu_time_us": 772.104, + "cuda_time_us": 219.69, + "pct_cuda_time": 1.1776982579350557, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 89.694, - "cuda_time_us": 51.516, - "pct_cuda_time": 0.2760924842294307, + "cpu_time_us": 91.106, + "cuda_time_us": 50.204, + "pct_cuda_time": 0.2691299710563591, "trace": "" }, "children": [ @@ -7646,8 +7646,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 51.516, - "pct_cuda_time": 0.2760924842294307, + "cuda_time_us": 50.204, + "pct_cuda_time": 0.2691299710563591, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -7657,9 +7657,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.012, - "cuda_time_us": 27.355, - "pct_cuda_time": 0.146605130563244, + "cpu_time_us": 80.12, + "cuda_time_us": 26.635, + "pct_cuda_time": 0.14278298101916434, "trace": "" }, "children": [ @@ -7667,8 +7667,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.025549227990889323, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -7677,8 +7677,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.565, - "pct_cuda_time": 0.024465451325944392, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -7687,8 +7687,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.37, - "pct_cuda_time": 0.0394984394900789, + "cuda_time_us": 7.33, + "pct_cuda_time": 0.039294133691401334, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -7697,8 +7697,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.848, - "pct_cuda_time": 0.03134150259674103, + "cuda_time_us": 5.847, + "pct_cuda_time": 0.03134417458303187, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -7707,8 +7707,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.02597142981939245, + "cuda_time_us": 4.126, + "pct_cuda_time": 0.02211836229341363, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -7718,9 +7718,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 448.309, - "cuda_time_us": 98.185, - "pct_cuda_time": 0.5262081792853998, + "cpu_time_us": 462.441, + "cuda_time_us": 100.068, + "pct_cuda_time": 0.5364372947109344, "trace": "" }, "children": [ @@ -7728,8 +7728,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.768, - "pct_cuda_time": 0.030912754271204216, + "cuda_time_us": 6.489, + "pct_cuda_time": 0.03478576173581217, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -7738,8 +7738,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 92.417, - "pct_cuda_time": 0.49529542501419555, + "cuda_time_us": 93.579, + "pct_cuda_time": 0.5016515329751221, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -7749,9 +7749,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.43, - "cuda_time_us": 40.98, - "pct_cuda_time": 0.21962632975623245, + "cpu_time_us": 101.161, + "cuda_time_us": 42.783, + "pct_cuda_time": 0.22934801114859799, "trace": "" }, "children": [ @@ -7759,8 +7759,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 40.98, - "pct_cuda_time": 0.21962632975623245, + "cuda_time_us": 42.783, + "pct_cuda_time": 0.22934801114859799, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -7772,9 +7772,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 39.098, + "cpu_time_us": 38.458, "cuda_time_us": 10.535, - "pct_cuda_time": 0.05646079511912906, + "pct_cuda_time": 0.05647526581704134, "trace": "" }, "children": [ @@ -7783,7 +7783,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 10.535, - "pct_cuda_time": 0.05646079511912906, + "pct_cuda_time": 0.05647526581704134, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7793,18 +7793,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 245.116, - "cuda_time_us": 313.964, - "pct_cuda_time": 1.6826442409854994, + "cpu_time_us": 222.262, + "cuda_time_us": 310.78, + "pct_cuda_time": 1.6660069397835886, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 86.549, - "cuda_time_us": 186.267, - "pct_cuda_time": 0.9982708044095694, + "cpu_time_us": 85.268, + "cuda_time_us": 182.772, + "pct_cuda_time": 0.9797909144672311, "trace": "" }, "children": [ @@ -7812,8 +7812,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 186.267, - "pct_cuda_time": 0.9982708044095694, + "cuda_time_us": 182.772, + "pct_cuda_time": 0.9797909144672311, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -7823,9 +7823,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 42.523, - "cuda_time_us": 13.299, - "pct_cuda_time": 0.07127404976642596, + "cpu_time_us": 27.061, + "cuda_time_us": 13.219, + "pct_cuda_time": 0.07086345883583005, "trace": "" }, "children": [ @@ -7833,8 +7833,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.299, - "pct_cuda_time": 0.07127404976642596, + "cuda_time_us": 13.219, + "pct_cuda_time": 0.07086345883583005, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -7844,9 +7844,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 97.075, - "cuda_time_us": 114.39800000000001, - "pct_cuda_time": 0.6130993868095043, + "cpu_time_us": 91.998, + "cuda_time_us": 114.78899999999999, + "pct_cuda_time": 0.6153525664805276, "trace": "" }, "children": [ @@ -7854,8 +7854,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 107.79, - "pct_cuda_time": 0.5776847751201635, + "cuda_time_us": 108.341, + "pct_cuda_time": 0.5807865945784599, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7864,8 +7864,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.608, - "pct_cuda_time": 0.03541461168934075, + "cuda_time_us": 6.448, + "pct_cuda_time": 0.03456597190206764, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -7879,18 +7879,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1061.775, - "cuda_time_us": 549.547, - "pct_cuda_time": 2.945216950672237, + "cpu_time_us": 1075.809, + "cuda_time_us": 549.2270000000001, + "pct_cuda_time": 2.9442563662929446, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 36.004, - "cuda_time_us": 10.295, - "pct_cuda_time": 0.05517455014251862, + "cpu_time_us": 35.733, + "cuda_time_us": 9.734, + "pct_cuda_time": 0.052181322967544415, "trace": "" }, "children": [ @@ -7898,8 +7898,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.295, - "pct_cuda_time": 0.05517455014251862, + "cuda_time_us": 9.734, + "pct_cuda_time": 0.052181322967544415, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -7909,18 +7909,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 744.201, - "cuda_time_us": 218.11700000000002, - "pct_cuda_time": 1.168966231513913, + "cpu_time_us": 760.316, + "cuda_time_us": 219.97000000000003, + "pct_cuda_time": 1.1791992616777016, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 86.379, - "cuda_time_us": 51.796, - "pct_cuda_time": 0.2775931033688096, + "cpu_time_us": 91.437, + "cuda_time_us": 51.606, + "pct_cuda_time": 0.27664571122489184, "trace": "" }, "children": [ @@ -7928,8 +7928,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 51.796, - "pct_cuda_time": 0.2775931033688096, + "cuda_time_us": 51.606, + "pct_cuda_time": 0.27664571122489184, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -7939,9 +7939,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 77.876, - "cuda_time_us": 26.755, - "pct_cuda_time": 0.1433895181217179, + "cpu_time_us": 85.808, + "cuda_time_us": 26.314, + "pct_cuda_time": 0.14106218744277418, "trace": "" }, "children": [ @@ -7949,8 +7949,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.926, - "pct_cuda_time": 0.026400178144929263, + "cuda_time_us": 4.806, + "pct_cuda_time": 0.025763657096981554, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -7959,8 +7959,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.0246915115665204, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -7969,8 +7969,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 6.689, - "pct_cuda_time": 0.03584871936894678, + "cuda_time_us": 7.049, + "pct_cuda_time": 0.03778776922110341, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -7979,8 +7979,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 6.289, - "pct_cuda_time": 0.03370497774126271, + "cuda_time_us": 5.728, + "pct_cuda_time": 0.03070624799240748, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -7989,8 +7989,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.285, - "pct_cuda_time": 0.022964832186565545, + "cuda_time_us": 4.125, + "pct_cuda_time": 0.022113001565761323, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -8000,9 +8000,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 442.641, - "cuda_time_us": 98.265, - "pct_cuda_time": 0.5266369276109366, + "cpu_time_us": 439.357, + "cuda_time_us": 100.108, + "pct_cuda_time": 0.5366517238170266, "trace": "" }, "children": [ @@ -8010,8 +8010,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.928, - "pct_cuda_time": 0.031770250922277844, + "cuda_time_us": 6.129, + "pct_cuda_time": 0.03285589978098209, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8020,8 +8020,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 92.337, - "pct_cuda_time": 0.4948666766886588, + "cuda_time_us": 93.979, + "pct_cuda_time": 0.5037958240360445, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8031,9 +8031,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.801, - "cuda_time_us": 41.301, - "pct_cuda_time": 0.22134668241244895, + "cpu_time_us": 105.157, + "cuda_time_us": 41.942, + "pct_cuda_time": 0.22483963919300884, "trace": "" }, "children": [ @@ -8041,8 +8041,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 41.301, - "pct_cuda_time": 0.22134668241244895, + "cuda_time_us": 41.942, + "pct_cuda_time": 0.22483963919300884, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -8054,9 +8054,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.217, - "cuda_time_us": 9.974, - "pct_cuda_time": 0.05345419748630216, + "cpu_time_us": 37.877, + "cuda_time_us": 10.535, + "pct_cuda_time": 0.05647526581704134, "trace": "" }, "children": [ @@ -8064,8 +8064,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.974, - "pct_cuda_time": 0.05345419748630216, + "cuda_time_us": 10.535, + "pct_cuda_time": 0.05647526581704134, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8075,18 +8075,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 221.771, - "cuda_time_us": 311.16100000000006, - "pct_cuda_time": 1.6676219715295038, + "cpu_time_us": 219.338, + "cuda_time_us": 308.988, + "pct_cuda_time": 1.656400515830657, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.718, - "cuda_time_us": 182.942, - "pct_cuda_time": 0.9804509521294456, + "cpu_time_us": 83.776, + "cuda_time_us": 182.001, + "pct_cuda_time": 0.9756577934473034, "trace": "" }, "children": [ @@ -8094,8 +8094,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 182.942, - "pct_cuda_time": 0.9804509521294456, + "cuda_time_us": 182.001, + "pct_cuda_time": 0.9756577934473034, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -8105,9 +8105,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 27.921, - "cuda_time_us": 13.419, - "pct_cuda_time": 0.07191717225473117, + "cpu_time_us": 27.26, + "cuda_time_us": 14.541, + "pct_cuda_time": 0.07795034079217827, "trace": "" }, "children": [ @@ -8115,8 +8115,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.419, - "pct_cuda_time": 0.07191717225473117, + "cuda_time_us": 14.541, + "pct_cuda_time": 0.07795034079217827, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -8126,9 +8126,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 89.454, - "cuda_time_us": 114.80000000000001, - "pct_cuda_time": 0.6152538471453267, + "cpu_time_us": 90.775, + "cuda_time_us": 112.44600000000001, + "pct_cuda_time": 0.6027923815911752, "trace": "" }, "children": [ @@ -8136,8 +8136,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 108.311, - "pct_cuda_time": 0.580476998590222, + "cuda_time_us": 105.677, + "pct_cuda_time": 0.5665056161127174, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8146,8 +8146,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.489, - "pct_cuda_time": 0.034776848555104746, + "cuda_time_us": 6.769, + "pct_cuda_time": 0.03628676547845779, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8161,18 +8161,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1045.05, - "cuda_time_us": 556.697, - "pct_cuda_time": 2.98353633226709, + "cpu_time_us": 1063.459, + "cuda_time_us": 547.472, + "pct_cuda_time": 2.934848289263147, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.563, + "cpu_time_us": 40.11, "cuda_time_us": 10.054, - "pct_cuda_time": 0.05388294581183898, + "pct_cuda_time": 0.05389675581628226, "trace": "" }, "children": [ @@ -8181,7 +8181,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 10.054, - "pct_cuda_time": 0.05388294581183898, + "pct_cuda_time": 0.05389675581628226, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8191,18 +8191,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 729.459, - "cuda_time_us": 220.10000000000002, - "pct_cuda_time": 1.1795938306331568, + "cpu_time_us": 735.059, + "cuda_time_us": 219.80800000000002, + "pct_cuda_time": 1.178330823798028, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 82.693, - "cuda_time_us": 50.654, - "pct_cuda_time": 0.27147272102177156, + "cpu_time_us": 86.98, + "cuda_time_us": 50.003, + "pct_cuda_time": 0.2680524647982457, "trace": "" }, "children": [ @@ -8210,8 +8210,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 50.654, - "pct_cuda_time": 0.27147272102177156, + "cuda_time_us": 50.003, + "pct_cuda_time": 0.2680524647982457, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -8221,9 +8221,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.232, - "cuda_time_us": 27.396, - "pct_cuda_time": 0.1468248640800816, + "cpu_time_us": 80.991, + "cuda_time_us": 27.035999999999998, + "pct_cuda_time": 0.1449326328077389, "trace": "" }, "children": [ @@ -8231,8 +8231,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.807, + "pct_cuda_time": 0.025769017824633863, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -8241,8 +8241,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.565, + "pct_cuda_time": 0.024471721732775863, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -8251,8 +8251,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.691, - "pct_cuda_time": 0.041218792146295355, + "cuda_time_us": 6.97, + "pct_cuda_time": 0.03736427173657125, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -8261,8 +8261,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.487, - "pct_cuda_time": 0.029406775777756165, + "cuda_time_us": 5.888, + "pct_cuda_time": 0.031563964416776405, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -8271,8 +8271,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.02532830733108723, + "cuda_time_us": 4.806, + "pct_cuda_time": 0.025763657096981554, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -8282,9 +8282,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 435.55, - "cuda_time_us": 99.06700000000001, - "pct_cuda_time": 0.5309351295744432, + "cpu_time_us": 431.896, + "cuda_time_us": 100.26700000000001, + "pct_cuda_time": 0.5375040795137431, "trace": "" }, "children": [ @@ -8292,8 +8292,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.688, - "pct_cuda_time": 0.030484005945667404, + "cuda_time_us": 5.727, + "pct_cuda_time": 0.030700887264755176, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8302,8 +8302,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 93.379, - "pct_cuda_time": 0.5004511236287759, + "cuda_time_us": 94.54, + "pct_cuda_time": 0.5068031922489881, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8313,9 +8313,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.771, - "cuda_time_us": 42.983, - "pct_cuda_time": 0.2303611159568604, + "cpu_time_us": 99.659, + "cuda_time_us": 42.502, + "pct_cuda_time": 0.2278416466783001, "trace": "" }, "children": [ @@ -8323,8 +8323,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.983, - "pct_cuda_time": 0.2303611159568604, + "cuda_time_us": 42.502, + "pct_cuda_time": 0.2278416466783001, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -8336,9 +8336,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.648, - "cuda_time_us": 10.495, - "pct_cuda_time": 0.056246420956360656, + "cpu_time_us": 45.488, + "cuda_time_us": 9.894, + "pct_cuda_time": 0.05303903939191334, "trace": "" }, "children": [ @@ -8346,8 +8346,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.495, - "pct_cuda_time": 0.056246420956360656, + "cuda_time_us": 9.894, + "pct_cuda_time": 0.05303903939191334, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8357,18 +8357,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 221.02, - "cuda_time_us": 316.048, - "pct_cuda_time": 1.6938131348657335, + "cpu_time_us": 220.48, + "cuda_time_us": 307.716, + "pct_cuda_time": 1.649581670256924, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 87.681, - "cuda_time_us": 188.891, - "pct_cuda_time": 1.012333749487177, + "cpu_time_us": 86.169, + "cuda_time_us": 180.219, + "pct_cuda_time": 0.9661049767708945, "trace": "" }, "children": [ @@ -8376,8 +8376,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 188.891, - "pct_cuda_time": 1.012333749487177, + "cuda_time_us": 180.219, + "pct_cuda_time": 0.9661049767708945, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -8387,9 +8387,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 25.929, - "cuda_time_us": 13.299, - "pct_cuda_time": 0.07127404976642596, + "cpu_time_us": 27.191, + "cuda_time_us": 13.339, + "pct_cuda_time": 0.07150674615410674, "trace": "" }, "children": [ @@ -8397,8 +8397,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.299, - "pct_cuda_time": 0.07127404976642596, + "cuda_time_us": 13.339, + "pct_cuda_time": 0.07150674615410674, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -8408,9 +8408,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 89.915, - "cuda_time_us": 113.858, - "pct_cuda_time": 0.6102053356121308, + "cpu_time_us": 88.362, + "cuda_time_us": 114.15799999999999, + "pct_cuda_time": 0.6119699473319226, "trace": "" }, "children": [ @@ -8418,8 +8418,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 107.089, - "pct_cuda_time": 0.573927867917647, + "cuda_time_us": 107.71, + "pct_cuda_time": 0.577403975429855, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8428,8 +8428,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.769, - "pct_cuda_time": 0.03627746769448359, + "cuda_time_us": 6.448, + "pct_cuda_time": 0.03456597190206764, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8443,18 +8443,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1044.239, - "cuda_time_us": 552.17, - "pct_cuda_time": 2.959274536395775, + "cpu_time_us": 1068.607, + "cuda_time_us": 551.13, + "pct_cuda_time": 2.954457831015282, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.863, - "cuda_time_us": 10.254, - "pct_cuda_time": 0.054954816625681004, + "cpu_time_us": 33.6, + "cuda_time_us": 9.854, + "pct_cuda_time": 0.0528246102858211, "trace": "" }, "children": [ @@ -8462,8 +8462,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.254, - "pct_cuda_time": 0.054954816625681004, + "cuda_time_us": 9.854, + "pct_cuda_time": 0.0528246102858211, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8473,18 +8473,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 732.324, - "cuda_time_us": 221.51299999999998, - "pct_cuda_time": 1.1871665979329507, + "cpu_time_us": 742.359, + "cuda_time_us": 219.56, + "pct_cuda_time": 1.177001363340256, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 84.106, - "cuda_time_us": 51.747, - "pct_cuda_time": 0.27733049501941825, + "cpu_time_us": 82.454, + "cuda_time_us": 50.595, + "pct_cuda_time": 0.2712260155684107, "trace": "" }, "children": [ @@ -8492,8 +8492,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 51.747, - "pct_cuda_time": 0.27733049501941825, + "cuda_time_us": 50.595, + "pct_cuda_time": 0.2712260155684107, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -8503,9 +8503,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 74.191, - "cuda_time_us": 27.836, - "pct_cuda_time": 0.14918297987053405, + "cpu_time_us": 74.722, + "cuda_time_us": 28.035999999999998, + "pct_cuda_time": 0.1502933604600447, "trace": "" }, "children": [ @@ -8513,8 +8513,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -8523,8 +8523,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.024899559005550416, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -8533,8 +8533,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.37, - "pct_cuda_time": 0.0394984394900789, + "cuda_time_us": 7.65, + "pct_cuda_time": 0.04100956654013918, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -8543,8 +8543,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 6.008, - "pct_cuda_time": 0.03219899924781466, + "cuda_time_us": 6.128, + "pct_cuda_time": 0.03285053905332979, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -8553,8 +8553,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.006, - "pct_cuda_time": 0.026828926470466075, + "cuda_time_us": 4.846, + "pct_cuda_time": 0.02597808620307379, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -8564,9 +8564,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 437.453, - "cuda_time_us": 99.427, - "pct_cuda_time": 0.5328644970393589, + "cpu_time_us": 447.469, + "cuda_time_us": 100.94999999999999, + "pct_cuda_time": 0.5411654565002679, "trace": "" }, "children": [ @@ -8574,8 +8574,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.768, - "pct_cuda_time": 0.030912754271204216, + "cuda_time_us": 5.728, + "pct_cuda_time": 0.03070624799240748, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8584,8 +8584,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 93.659, - "pct_cuda_time": 0.5019517427681547, + "cuda_time_us": 95.222, + "pct_cuda_time": 0.5104592085078604, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8595,9 +8595,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.691, - "cuda_time_us": 42.503, - "pct_cuda_time": 0.22778862600363955, + "cpu_time_us": 101.172, + "cuda_time_us": 39.979, + "pct_cuda_time": 0.21431653081153257, "trace": "" }, "children": [ @@ -8605,8 +8605,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.503, - "pct_cuda_time": 0.22778862600363955, + "cuda_time_us": 39.979, + "pct_cuda_time": 0.21431653081153257, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -8618,9 +8618,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 36.905, - "cuda_time_us": 10.174, - "pct_cuda_time": 0.054526068300144184, + "cpu_time_us": 37.686, + "cuda_time_us": 10.334, + "pct_cuda_time": 0.05539775955892788, "trace": "" }, "children": [ @@ -8628,8 +8628,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.174, - "pct_cuda_time": 0.054526068300144184, + "cuda_time_us": 10.334, + "pct_cuda_time": 0.05539775955892788, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8639,18 +8639,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 218.316, - "cuda_time_us": 310.229, - "pct_cuda_time": 1.6626270535369994, + "cpu_time_us": 232.467, + "cuda_time_us": 311.382, + "pct_cuda_time": 1.6692340978302769, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 84.767, - "cuda_time_us": 182.001, - "pct_cuda_time": 0.9754077999503188, + "cpu_time_us": 91.797, + "cuda_time_us": 182.953, + "pct_cuda_time": 0.9807612061722984, "trace": "" }, "children": [ @@ -8658,8 +8658,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 182.001, - "pct_cuda_time": 0.9754077999503188, + "cuda_time_us": 182.953, + "pct_cuda_time": 0.9807612061722984, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -8669,9 +8669,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.75, - "cuda_time_us": 13.659, - "pct_cuda_time": 0.07320341723134162, + "cpu_time_us": 27.801, + "cuda_time_us": 13.419, + "pct_cuda_time": 0.0719356043662912, "trace": "" }, "children": [ @@ -8679,8 +8679,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.659, - "pct_cuda_time": 0.07320341723134162, + "cuda_time_us": 13.419, + "pct_cuda_time": 0.0719356043662912, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -8690,9 +8690,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 89.484, - "cuda_time_us": 114.569, - "pct_cuda_time": 0.6140158363553391, + "cpu_time_us": 92.789, + "cuda_time_us": 115.01, + "pct_cuda_time": 0.6165372872916872, "trace": "" }, "children": [ @@ -8700,8 +8700,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 108.0, - "pct_cuda_time": 0.5788102394746976, + "cuda_time_us": 108.04, + "pct_cuda_time": 0.579173015555116, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8710,8 +8710,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.569, - "pct_cuda_time": 0.03520559688064156, + "cuda_time_us": 6.97, + "pct_cuda_time": 0.03736427173657125, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8725,18 +8725,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1067.593, - "cuda_time_us": 551.428, - "pct_cuda_time": 2.9552978956764213, + "cpu_time_us": 1196.178, + "cuda_time_us": 541.444, + "pct_cuda_time": 2.902533822975048, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 34.351, - "cuda_time_us": 10.094, - "pct_cuda_time": 0.05409731997460738, + "cpu_time_us": 34.622, + "cuda_time_us": 9.853, + "pct_cuda_time": 0.0528192495581688, "trace": "" }, "children": [ @@ -8744,8 +8744,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.094, - "pct_cuda_time": 0.05409731997460738, + "cuda_time_us": 9.853, + "pct_cuda_time": 0.0528192495581688, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8755,18 +8755,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 748.006, - "cuda_time_us": 219.389, - "pct_cuda_time": 1.1757833298899483, + "cpu_time_us": 853.926, + "cuda_time_us": 217.596, + "pct_cuda_time": 1.1664728942311275, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 81.932, - "cuda_time_us": 49.063, - "pct_cuda_time": 0.26294598869765823, + "cpu_time_us": 85.548, + "cuda_time_us": 49.393, + "pct_cuda_time": 0.26478242093033916, "trace": "" }, "children": [ @@ -8774,8 +8774,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 49.063, - "pct_cuda_time": 0.26294598869765823, + "cuda_time_us": 49.393, + "pct_cuda_time": 0.26478242093033916, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -8785,9 +8785,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 74.621, - "cuda_time_us": 27.916, - "pct_cuda_time": 0.1496117281960709, + "cpu_time_us": 83.335, + "cuda_time_us": 27.036, + "pct_cuda_time": 0.14493263280773894, "trace": "" }, "children": [ @@ -8795,8 +8795,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.026185803982160853, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -8805,8 +8805,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.025113933168318822, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.02490594067261263, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -8815,8 +8815,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.53, - "pct_cuda_time": 0.04035593614115252, + "cuda_time_us": 7.37, + "pct_cuda_time": 0.03950856279749357, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -8825,8 +8825,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.928, - "pct_cuda_time": 0.031770250922277844, + "cuda_time_us": 5.567, + "pct_cuda_time": 0.029843170840386252, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -8835,8 +8835,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.026185803982160853, + "cuda_time_us": 4.727, + "pct_cuda_time": 0.0253401596124494, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -8846,9 +8846,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 436.341, - "cuda_time_us": 99.547, - "pct_cuda_time": 0.533507619527664, + "cpu_time_us": 454.84, + "cuda_time_us": 99.947, + "pct_cuda_time": 0.5357886466650053, "trace": "" }, "children": [ @@ -8856,8 +8856,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.808, - "pct_cuda_time": 0.031127128433972626, + "cuda_time_us": 5.807, + "pct_cuda_time": 0.031129745476939637, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8866,8 +8866,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 93.739, - "pct_cuda_time": 0.5023804910936914, + "cuda_time_us": 94.14, + "pct_cuda_time": 0.5046589011880657, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -8877,9 +8877,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 120.931, - "cuda_time_us": 42.863, - "pct_cuda_time": 0.2297179934685552, + "cpu_time_us": 195.082, + "cuda_time_us": 41.22, + "pct_cuda_time": 0.22096919382804406, "trace": "" }, "children": [ @@ -8887,8 +8887,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 42.863, - "pct_cuda_time": 0.2297179934685552, + "cuda_time_us": 41.22, + "pct_cuda_time": 0.22096919382804406, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -8900,9 +8900,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 40.04, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cpu_time_us": 42.804, + "cuda_time_us": 10.415, + "pct_cuda_time": 0.055831978498764645, "trace": "" }, "children": [ @@ -8910,8 +8910,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.854, - "pct_cuda_time": 0.052811074997996935, + "cuda_time_us": 10.415, + "pct_cuda_time": 0.055831978498764645, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -8921,18 +8921,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 223.173, - "cuda_time_us": 312.091, - "pct_cuda_time": 1.6726061708138686, + "cpu_time_us": 241.16, + "cuda_time_us": 303.58, + "pct_cuda_time": 1.627409700686987, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.107, - "cuda_time_us": 184.734, - "pct_cuda_time": 0.9900549146214701, + "cpu_time_us": 91.036, + "cuda_time_us": 177.214, + "pct_cuda_time": 0.9499959901757156, "trace": "" }, "children": [ @@ -8940,8 +8940,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 184.734, - "pct_cuda_time": 0.9900549146214701, + "cuda_time_us": 177.214, + "pct_cuda_time": 0.9499959901757156, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -8951,9 +8951,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.64, - "cuda_time_us": 13.5, - "pct_cuda_time": 0.0723512799343372, + "cpu_time_us": 26.95, + "cuda_time_us": 13.539, + "pct_cuda_time": 0.0725788916845679, "trace": "" }, "children": [ @@ -8961,8 +8961,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.5, - "pct_cuda_time": 0.0723512799343372, + "cuda_time_us": 13.539, + "pct_cuda_time": 0.0725788916845679, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -8972,9 +8972,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 93.349, - "cuda_time_us": 113.857, - "pct_cuda_time": 0.6101999762580614, + "cpu_time_us": 103.465, + "cuda_time_us": 112.827, + "pct_cuda_time": 0.6048348188267036, "trace": "" }, "children": [ @@ -8982,8 +8982,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 106.928, - "pct_cuda_time": 0.5730650119125043, + "cuda_time_us": 106.298, + "pct_cuda_time": 0.5698346279847993, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -8992,8 +8992,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.929, - "pct_cuda_time": 0.037134964345557214, + "cuda_time_us": 6.529, + "pct_cuda_time": 0.03500019084190441, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -9007,18 +9007,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1049.467, - "cuda_time_us": 552.862, - "pct_cuda_time": 2.9629832094116684, + "cpu_time_us": 1073.024, + "cuda_time_us": 547.865, + "pct_cuda_time": 2.9369550552305035, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.342, - "cuda_time_us": 10.295, - "pct_cuda_time": 0.05517455014251862, + "cpu_time_us": 34.311, + "cuda_time_us": 9.774, + "pct_cuda_time": 0.052395752073636635, "trace": "" }, "children": [ @@ -9026,8 +9026,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.295, - "pct_cuda_time": 0.05517455014251862, + "cuda_time_us": 9.774, + "pct_cuda_time": 0.052395752073636635, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -9037,18 +9037,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 737.111, - "cuda_time_us": 221.022, - "pct_cuda_time": 1.1845351550849683, + "cpu_time_us": 747.537, + "cuda_time_us": 221.24199999999996, + "pct_cuda_time": 1.1860181072514342, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 84.286, - "cuda_time_us": 50.374, - "pct_cuda_time": 0.26997210188239273, + "cpu_time_us": 83.635, + "cuda_time_us": 49.873, + "pct_cuda_time": 0.2673555702034459, "trace": "" }, "children": [ @@ -9056,8 +9056,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 50.374, - "pct_cuda_time": 0.26997210188239273, + "cuda_time_us": 49.873, + "pct_cuda_time": 0.2673555702034459, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -9067,9 +9067,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 72.628, - "cuda_time_us": 28.197, - "pct_cuda_time": 0.15111770668951896, + "cpu_time_us": 74.271, + "cuda_time_us": 26.154, + "pct_cuda_time": 0.14020447101840525, "trace": "" }, "children": [ @@ -9077,8 +9077,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.026185803982160853, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -9087,8 +9087,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -9097,8 +9097,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.531, - "pct_cuda_time": 0.04036129549522173, + "cuda_time_us": 7.13, + "pct_cuda_time": 0.038221988160940176, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -9107,8 +9107,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 6.008, - "pct_cuda_time": 0.03219899924781466, + "cuda_time_us": 5.567, + "pct_cuda_time": 0.029843170840386252, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -9117,8 +9117,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.006, - "pct_cuda_time": 0.026828926470466075, + "cuda_time_us": 4.165, + "pct_cuda_time": 0.022327430671853554, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -9128,9 +9128,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 440.988, - "cuda_time_us": 100.549, - "pct_cuda_time": 0.5388776923050126, + "cpu_time_us": 444.675, + "cuda_time_us": 102.11099999999999, + "pct_cuda_time": 0.547389261304595, "trace": "" }, "children": [ @@ -9138,8 +9138,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.608, - "pct_cuda_time": 0.03005525762013059, + "cuda_time_us": 6.609, + "pct_cuda_time": 0.03542904905408887, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -9148,8 +9148,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 94.941, - "pct_cuda_time": 0.508822434684882, + "cuda_time_us": 95.502, + "pct_cuda_time": 0.511960212250506, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -9159,9 +9159,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 102.934, - "cuda_time_us": 41.902, - "pct_cuda_time": 0.22456765420804425, + "cpu_time_us": 109.484, + "cuda_time_us": 43.104, + "pct_cuda_time": 0.23106880472498814, "trace": "" }, "children": [ @@ -9169,8 +9169,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 41.902, - "pct_cuda_time": 0.22456765420804425, + "cuda_time_us": 43.104, + "pct_cuda_time": 0.23106880472498814, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -9182,9 +9182,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.556, - "cuda_time_us": 10.414, - "pct_cuda_time": 0.05581231327675463, + "cpu_time_us": 40.11, + "cuda_time_us": 10.294, + "pct_cuda_time": 0.05518333045283566, "trace": "" }, "children": [ @@ -9192,8 +9192,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.414, - "pct_cuda_time": 0.05581231327675463, + "cuda_time_us": 10.294, + "pct_cuda_time": 0.05518333045283566, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -9203,18 +9203,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 218.046, - "cuda_time_us": 311.131, - "pct_cuda_time": 1.667461190907427, + "cpu_time_us": 227.88, + "cuda_time_us": 306.555, + "pct_cuda_time": 1.6433578654525969, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 84.376, - "cuda_time_us": 184.184, - "pct_cuda_time": 0.9871072698834046, + "cpu_time_us": 89.204, + "cuda_time_us": 178.987, + "pct_cuda_time": 0.9595005603032537, "trace": "" }, "children": [ @@ -9222,8 +9222,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 184.184, - "pct_cuda_time": 0.9871072698834046, + "cuda_time_us": 178.987, + "pct_cuda_time": 0.9595005603032537, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -9233,9 +9233,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.119, - "cuda_time_us": 13.339, - "pct_cuda_time": 0.07148842392919436, + "cpu_time_us": 27.071, + "cuda_time_us": 13.299, + "pct_cuda_time": 0.0712923170480145, "trace": "" }, "children": [ @@ -9243,8 +9243,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.339, - "pct_cuda_time": 0.07148842392919436, + "cuda_time_us": 13.299, + "pct_cuda_time": 0.0712923170480145, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -9254,9 +9254,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 88.932, - "cuda_time_us": 113.60799999999999, - "pct_cuda_time": 0.6088654970948281, + "cpu_time_us": 92.358, + "cuda_time_us": 114.269, + "pct_cuda_time": 0.6125649881013286, "trace": "" }, "children": [ @@ -9264,8 +9264,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 107.079, - "pct_cuda_time": 0.5738742743769549, + "cuda_time_us": 107.459, + "pct_cuda_time": 0.5760584327891263, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -9274,8 +9274,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.529, - "pct_cuda_time": 0.03499122271787315, + "cuda_time_us": 6.81, + "pct_cuda_time": 0.03650655531220232, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -9289,18 +9289,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 1044.649, - "cuda_time_us": 547.362, - "pct_cuda_time": 2.933506762031013, + "cpu_time_us": 1078.563, + "cuda_time_us": 550.1569999999999, + "pct_cuda_time": 2.9492418430095877, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.023, - "cuda_time_us": 10.014, - "pct_cuda_time": 0.05366857164907056, + "cpu_time_us": 42.904, + "cuda_time_us": 10.134, + "pct_cuda_time": 0.05432561402846672, "trace": "" }, "children": [ @@ -9308,8 +9308,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.014, - "pct_cuda_time": 0.05366857164907056, + "cuda_time_us": 10.134, + "pct_cuda_time": 0.05432561402846672, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -9319,18 +9319,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 730.841, - "cuda_time_us": 219.50900000000001, - "pct_cuda_time": 1.1764264523782537, + "cpu_time_us": 756.18, + "cuda_time_us": 221.191, + "pct_cuda_time": 1.1857447101411667, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 83.234, - "cuda_time_us": 51.706, - "pct_cuda_time": 0.2771107615025807, + "cpu_time_us": 85.979, + "cuda_time_us": 49.543, + "pct_cuda_time": 0.265586530078185, "trace": "" }, "children": [ @@ -9338,8 +9338,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x128x128_MI16x16x1_SN_LDSB1_AFC0_AFEM8_AFEM8_ASEM32_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_4_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 51.706, - "pct_cuda_time": 0.2771107615025807, + "cuda_time_us": 49.543, + "pct_cuda_time": 0.265586530078185, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 6144]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 6144]) <- linear(bfloat16[256, 4096], bfloat16[6144, 4096], None)" }, "children": [] @@ -9349,9 +9349,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 75.873, - "cuda_time_us": 27.275000000000002, - "pct_cuda_time": 0.14617638223770718, + "cpu_time_us": 81.712, + "cuda_time_us": 27.195, + "pct_cuda_time": 0.14578498850445556, "trace": "" }, "children": [ @@ -9359,8 +9359,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.026185803982160853, + "cuda_time_us": 4.926, + "pct_cuda_time": 0.026406944415258253, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False) <- clone(bfloat16[256, 4096], 0) <- contiguous(bfloat16[256, 4096], 0)" }, "children": [] @@ -9369,8 +9369,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.024685184842782006, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.02447708246042817, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False) <- clone(bfloat16[256, 1024], 0) <- contiguous(bfloat16[256, 1024], 0)" }, "children": [] @@ -9379,8 +9379,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 7.049, - "pct_cuda_time": 0.03777808683386244, + "cuda_time_us": 7.61, + "pct_cuda_time": 0.04079513743404695, "trace": "_C::rotary_embedding(int64[256], bfloat16[256, 4096], bfloat16[256, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -9389,8 +9389,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.928, - "pct_cuda_time": 0.031770250922277844, + "cuda_time_us": 5.968, + "pct_cuda_time": 0.03199282262896087, "trace": "copy_(bfloat16[256, 4096], bfloat16[256, 4096], False)" }, "children": [] @@ -9399,8 +9399,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl_nocast(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#12}::operator()() const::{lambda(c10::BFloat16)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.125, + "pct_cuda_time": 0.022113001565761323, "trace": "copy_(bfloat16[256, 1024], bfloat16[256, 1024], False)" }, "children": [] @@ -9410,9 +9410,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 436.081, - "cuda_time_us": 100.74900000000001, - "pct_cuda_time": 0.5399495631188548, + "cpu_time_us": 449.442, + "cuda_time_us": 100.429, + "pct_cuda_time": 0.5383725173934167, "trace": "" }, "children": [ @@ -9420,8 +9420,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.968, - "pct_cuda_time": 0.03198462508504625, + "cuda_time_us": 6.409, + "pct_cuda_time": 0.03435690352362771, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[256], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -9430,8 +9430,8 @@ "entry": { "name": "attn_fwd", "cpu_time_us": 0, - "cuda_time_us": 94.781, - "pct_cuda_time": 0.5079649380338084, + "cuda_time_us": 94.02, + "pct_cuda_time": 0.504015613869789, "trace": "_attention(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], int32[2], int32[2], 256, 256, True, 0.08838834764831845, None, None, None) <- vllm::unified_attention_with_output(bfloat16[256, 32, 128], bfloat16[256, 8, 128], bfloat16[256, 8, 128], bfloat16[256, 32, 128], None, None)" }, "children": [] @@ -9441,9 +9441,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 100.39, - "cuda_time_us": 39.779, - "pct_cuda_time": 0.21318974551911107, + "cpu_time_us": 102.273, + "cuda_time_us": 44.024, + "pct_cuda_time": 0.23600067416510948, "trace": "" }, "children": [ @@ -9451,8 +9451,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT64x64x128_MI16x16x1_SN_LDSB1_AFC0_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA512_LBSPPB256_LBSPPM0_LPA16_LPB16_LPMn1_LRVW8_LWPMn1_MIAV0_MIWT4_2_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS0_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB1_WSGRA0_WSGRB0_WS64_WG16_8_2", "cpu_time_us": 0, - "cuda_time_us": 39.779, - "pct_cuda_time": 0.21318974551911107, + "cuda_time_us": 44.024, + "pct_cuda_time": 0.23600067416510948, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 4096]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 4096]) <- linear(bfloat16[256, 4096], bfloat16[4096, 4096], None)" }, "children": [] @@ -9464,9 +9464,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 38.147, - "cuda_time_us": 10.054, - "pct_cuda_time": 0.05388294581183898, + "cpu_time_us": 38.288, + "cuda_time_us": 10.255, + "pct_cuda_time": 0.05497426207439573, "trace": "" }, "children": [ @@ -9474,8 +9474,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.054, - "pct_cuda_time": 0.05388294581183898, + "cuda_time_us": 10.255, + "pct_cuda_time": 0.05497426207439573, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -9485,18 +9485,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 219.478, - "cuda_time_us": 307.78499999999997, - "pct_cuda_time": 1.6495287921918498, + "cpu_time_us": 218.777, + "cuda_time_us": 308.577, + "pct_cuda_time": 1.6541972567655592, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 85.207, - "cuda_time_us": 179.016, - "pct_cuda_time": 0.9594101280537264, + "cpu_time_us": 83.746, + "cuda_time_us": 181.109, + "pct_cuda_time": 0.9708760243814467, "trace": "" }, "children": [ @@ -9504,8 +9504,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT96x256x64_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA256_LBSPPB512_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT6_4_MO40_NTn1_NTA4_NTB0_NTC0_NTD4_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW2_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA2_VWB4_WSGRA0_WSGRB0_WS64_WG16_16_1", "cpu_time_us": 0, - "cuda_time_us": 179.016, - "pct_cuda_time": 0.9594101280537264, + "cuda_time_us": 181.109, + "pct_cuda_time": 0.9708760243814467, "trace": "mm(bfloat16[256, 4096], bfloat16[4096, 28672]) <- matmul(bfloat16[256, 4096], bfloat16[4096, 28672]) <- linear(bfloat16[256, 4096], bfloat16[28672, 4096], None)" }, "children": [] @@ -9515,9 +9515,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 26.329, - "cuda_time_us": 13.579, - "pct_cuda_time": 0.07277466890580479, + "cpu_time_us": 27.02, + "cuda_time_us": 14.581, + "pct_cuda_time": 0.07816476989827051, "trace": "" }, "children": [ @@ -9525,8 +9525,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 13.579, - "pct_cuda_time": 0.07277466890580479, + "cuda_time_us": 14.581, + "pct_cuda_time": 0.07816476989827051, "trace": "_C::silu_and_mul(bfloat16[256, 14336], bfloat16[256, 28672])" }, "children": [] @@ -9536,9 +9536,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 90.956, - "cuda_time_us": 115.19, - "pct_cuda_time": 0.6173439952323185, + "cpu_time_us": 89.964, + "cuda_time_us": 112.887, + "pct_cuda_time": 0.6051564624858421, "trace": "" }, "children": [ @@ -9546,8 +9546,8 @@ "entry": { "name": "Cijk_Alik_Bljk_BBS_BH_Bias_HAS_SAV_UserArgs_MT128x96x128_MI16x16x1_SN_LDSB1_AFC1_AFEM1_AFEM1_ASEM1_CLR1_CADS0_DTVA0_DTVB0_EPS0_FDSI0_GRPM1_GRVWA8_GRVWB8_GSUAMB_GLS0_ISA942_IU1_K1_LBSPPA1024_LBSPPB256_LBSPPM0_LPA16_LPB16_LPM0_LRVW8_LWPMn1_MIAV0_MIWT4_3_MO40_NTn1_NTA0_NTB0_NTC0_NTD0_NTM0_NEPBS16_NLCA1_NLCB1_ONLL1_PGR2_PLR1_PKA1_SIA3_SS1_SPO0_SRVW0_SSO0_SVW4_SK0_SKXCCM0_TLDS1_ULSGRO0_USL1_UIOFGRO0_USFGROn1_VSn1_VWA4_VWB1_WSGRA0_WSGRB0_WS64_WG32_8_1", "cpu_time_us": 0, - "cuda_time_us": 108.301, - "pct_cuda_time": 0.5804234050495298, + "cuda_time_us": 106.839, + "pct_cuda_time": 0.5727347816446967, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -9556,8 +9556,8 @@ "entry": { "name": "Cijk_SB_BiasS_HAS_ScaleAlphaVec_PostGSU4_VW4", "cpu_time_us": 0, - "cuda_time_us": 6.889, - "pct_cuda_time": 0.03692059018278881, + "cuda_time_us": 6.048, + "pct_cuda_time": 0.03242168084114533, "trace": "mm(bfloat16[256, 14336], bfloat16[14336, 4096]) <- matmul(bfloat16[256, 14336], bfloat16[14336, 4096]) <- linear(bfloat16[256, 14336], bfloat16[4096, 14336], None)" }, "children": [] @@ -9571,9 +9571,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.943, - "cuda_time_us": 9.974, - "pct_cuda_time": 0.05345419748630216, + "cpu_time_us": 35.123, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "" }, "children": [ @@ -9581,8 +9581,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 9.974, - "pct_cuda_time": 0.05345419748630216, + "cuda_time_us": 9.934, + "pct_cuda_time": 0.053253468498005566, "trace": "_C::fused_add_rms_norm(bfloat16[256, 4096], bfloat16[256, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -9594,9 +9594,9 @@ { "entry": { "name": "LogitsProcessor", - "cpu_time_us": 142.503, - "cuda_time_us": 241.549, - "pct_cuda_time": 1.2945466160636456, + "cpu_time_us": 144.676, + "cuda_time_us": 241.62900000000002, + "pct_cuda_time": 1.2953072618989923, "trace": "" }, "children": [ @@ -9604,8 +9604,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::indexSelectSmallIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, long)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.0244708106800136, + "cuda_time_us": 4.085, + "pct_cuda_time": 0.021898572459669093, "trace": "index_select(bfloat16[256, 4096], 0, int64[1])" }, "children": [] @@ -9614,8 +9614,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 236.983, - "pct_cuda_time": 1.270075805383632, + "cuda_time_us": 237.544, + "pct_cuda_time": 1.2734086894393233, "trace": "_rocm_C::wvSplitK(bfloat16[128256, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -9625,9 +9625,9 @@ { "entry": { "name": "Sampler", - "cpu_time_us": 804.01, - "cuda_time_us": 226.40500000000003, - "pct_cuda_time": 1.213384558039527, + "cpu_time_us": 793.887, + "cuda_time_us": 222.557, + "pct_cuda_time": 1.1930674641142165, "trace": "" }, "children": [ @@ -9635,8 +9635,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.02575705565662404, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.025334798884797092, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -9645,8 +9645,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.02597142981939245, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -9655,8 +9655,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(int32[1], int32[1], True) <- _to_copy(int32[1], 3, 0, None, None, True, None) <- to(int32[1], 3, 0, None, None, True, False, None)" }, "children": [] @@ -9665,8 +9665,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.927, - "pct_cuda_time": 0.026405537498998467, + "cuda_time_us": 4.767, + "pct_cuda_time": 0.025554588718541632, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -9675,8 +9675,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 5.006, - "pct_cuda_time": 0.026828926470466075, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.025549227990889323, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -9685,8 +9685,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.927, - "pct_cuda_time": 0.026405537498998467, + "cuda_time_us": 3.884, + "pct_cuda_time": 0.02082106620155563, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -9695,8 +9695,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.165, - "pct_cuda_time": 0.022321709698260327, + "cuda_time_us": 2.442, + "pct_cuda_time": 0.013090896926930704, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -9705,8 +9705,8 @@ "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 2.443, - "pct_cuda_time": 0.013092901991080427, + "cuda_time_us": 3.524, + "pct_cuda_time": 0.01889120424672555, "trace": "copy_(float32[1, 128256], bfloat16[1, 128256], False) <- _to_copy(bfloat16[1, 128256], 6, None, None, None, False, None) <- to(bfloat16[1, 128256], 6, False, False, None)" }, "children": [] @@ -9715,8 +9715,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2}>(int, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2})", "cpu_time_us": 0, - "cuda_time_us": 14.141, - "pct_cuda_time": 0.07578662589270091, + "cuda_time_us": 13.94, + "pct_cuda_time": 0.0747285434731425, "trace": "div_(float32[1, 128256], bfloat16[1, 1])" }, "children": [] @@ -9725,8 +9725,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::SoftMaxForwardEpilogue>(float*, float const*, int)", "cpu_time_us": 0, - "cuda_time_us": 70.985, - "pct_cuda_time": 0.3804337486028834, + "cuda_time_us": 71.145, + "pct_cuda_time": 0.38138896882329437, "trace": "_softmax(float32[1, 128256], -1, False) <- softmax(float32[1, 128256], -1, 6)" }, "children": [] @@ -9735,8 +9735,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::LogSoftMaxForwardEpilogue>(float*, float const*, int)", "cpu_time_us": 0, - "cuda_time_us": 40.42, - "pct_cuda_time": 0.2166250914774748, + "cuda_time_us": 40.259, + "pct_cuda_time": 0.21581753455417824, "trace": "_log_softmax(float32[1, 128256], -1, False) <- log_softmax(float32[1, 128256], -1, 6)" }, "children": [] @@ -9745,8 +9745,8 @@ "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.025542681493855638, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(int64[1], int32[1], False) <- _to_copy(int32[1], 4, None, None, None, False, None) <- to(int32[1], 4, False, False, None)" }, "children": [] @@ -9755,8 +9755,8 @@ "entry": { "name": "void at::native::index_elementwise_kernel<128, 4, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1}>(long, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 8.251, - "pct_cuda_time": 0.04422003042505305, + "cuda_time_us": 6.73, + "pct_cuda_time": 0.03607769710001787, "trace": "index(float32[1, 128256], None)" }, "children": [] @@ -9765,8 +9765,8 @@ "entry": { "name": "void at::native::reduce_kernel<512, 1, at::native::ReduceOp, unsigned int, long, 4, 4> >(at::native::ReduceOp, unsigned int, long, 4, 4>)", "cpu_time_us": 0, - "cuda_time_us": 46.949, - "pct_cuda_time": 0.2516163141953479, + "cuda_time_us": 47.63, + "pct_cuda_time": 0.2553314580793241, "trace": "argmax(float32[1, 128256], -1, False)" }, "children": [] @@ -9775,8 +9775,8 @@ "entry": { "name": "Memcpy DtoH (Device -> Host)", "cpu_time_us": 0, - "cuda_time_us": 5.007, - "pct_cuda_time": 0.02683428582453528, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.02512036977870486, "trace": "copy_(int64[1], int64[1], False) <- _to_copy(int64[1], 4, 0, None, None, False, None) <- to(int64[1], 4, 0, None, None, False, False, None)" }, "children": [] @@ -9793,24 +9793,24 @@ { "entry": { "name": "LlamaForCausalLM", - "cuda_time_us": 5292.264999999999, - "pct_cuda_time": 92.2104376570522, + "cuda_time_us": 5316.416, + "pct_cuda_time": 92.18150150806221, "invocations": 1 }, "children": [ { "entry": { "name": "VocabParallelEmbedding(weight=bfloat16[128256, 4096])", - "cuda_time_us": 3.844, - "pct_cuda_time": 0.0669764122457414, + "cuda_time_us": 4.165, + "pct_cuda_time": 0.0722170638605179, "invocations": 1 }, "children": [ { "entry": { "name": "void at::native::(anonymous namespace)::indexSelectSmallIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, long)", - "cuda_time_us": 3.844, - "pct_cuda_time": 0.0669764122457414, + "cuda_time_us": 4.165, + "pct_cuda_time": 0.0722170638605179, "invocations": 1 }, "children": [] @@ -9820,24 +9820,24 @@ { "entry": { "name": "LlamaDecoderLayer", - "cuda_time_us": 5284.095, - "pct_cuda_time": 92.06808664559338, + "cuda_time_us": 5307.685000000001, + "pct_cuda_time": 92.03011442893471, "invocations": 32 }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cuda_time_us": 304.3069999999999, - "pct_cuda_time": 5.302130874418529, + "cuda_time_us": 307.035, + "pct_cuda_time": 5.3236893643251175, "invocations": 64 }, "children": [ { "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::rms_norm_kernel(c10::BFloat16*, c10::BFloat16 const*, c10::BFloat16 const*, float, int, unsigned long, unsigned long)", - "cuda_time_us": 5.207, - "pct_cuda_time": 0.09072481232142963, + "cuda_time_us": 5.206, + "pct_cuda_time": 0.09026699506791266, "invocations": 1 }, "children": [] @@ -9845,8 +9845,8 @@ { "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", - "cuda_time_us": 299.0999999999999, - "pct_cuda_time": 5.211406062097099, + "cuda_time_us": 301.829, + "pct_cuda_time": 5.233422369257204, "invocations": 63 }, "children": [] @@ -9856,24 +9856,24 @@ { "entry": { "name": "LlamaAttention", - "cuda_time_us": 1559.3070000000002, - "pct_cuda_time": 27.16877951344181, + "cuda_time_us": 1574.7330000000002, + "pct_cuda_time": 27.30434420750659, "invocations": 32 }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cuda_time_us": 564.5269999999999, - "pct_cuda_time": 9.836106419316247, + "cuda_time_us": 554.549, + "pct_cuda_time": 9.615342268135976, "invocations": 32 }, "children": [ { "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", - "cuda_time_us": 564.5269999999999, - "pct_cuda_time": 9.836106419316247, + "cuda_time_us": 554.549, + "pct_cuda_time": 9.615342268135976, "invocations": 32 }, "children": [] @@ -9883,16 +9883,16 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cuda_time_us": 125.679, - "pct_cuda_time": 2.1897836926723557, + "cuda_time_us": 123.83199999999998, + "pct_cuda_time": 2.1471268792258464, "invocations": 32 }, "children": [ { "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", - "cuda_time_us": 125.679, - "pct_cuda_time": 2.1897836926723557, + "cuda_time_us": 123.83199999999998, + "pct_cuda_time": 2.1471268792258464, "invocations": 32 }, "children": [] @@ -9902,16 +9902,16 @@ { "entry": { "name": "Attention", - "cuda_time_us": 423.3379999999999, - "pct_cuda_time": 7.376082311989508, + "cuda_time_us": 463.03, + "pct_cuda_time": 8.0284914956388, "invocations": 32 }, "children": [ { "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", - "cuda_time_us": 89.26100000000002, - "pct_cuda_time": 1.5552501387791688, + "cuda_time_us": 92.22399999999999, + "pct_cuda_time": 1.5990747893091, "invocations": 32 }, "children": [] @@ -9919,8 +9919,8 @@ { "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", - "cuda_time_us": 204.47299999999998, - "pct_cuda_time": 3.5626607547147455, + "cuda_time_us": 249.53600000000003, + "pct_cuda_time": 4.3267124243693145, "invocations": 32 }, "children": [] @@ -9928,8 +9928,8 @@ { "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", - "cuda_time_us": 129.60399999999998, - "pct_cuda_time": 2.2581714184955954, + "cuda_time_us": 121.27000000000001, + "pct_cuda_time": 2.102704281960386, "invocations": 32 }, "children": [] @@ -9939,16 +9939,16 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cuda_time_us": 445.7629999999999, - "pct_cuda_time": 7.76680708946369, + "cuda_time_us": 433.322, + "pct_cuda_time": 7.513383564505963, "invocations": 32 }, "children": [ { "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", - "cuda_time_us": 445.7629999999999, - "pct_cuda_time": 7.76680708946369, + "cuda_time_us": 433.322, + "pct_cuda_time": 7.513383564505963, "invocations": 32 }, "children": [] @@ -9960,24 +9960,24 @@ { "entry": { "name": "LlamaMLP", - "cuda_time_us": 3420.481, - "pct_cuda_time": 59.59717625773304, + "cuda_time_us": 3425.9170000000004, + "pct_cuda_time": 59.40208085710298, "invocations": 32 }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cuda_time_us": 2051.6830000000004, - "pct_cuda_time": 35.74775400769498, + "cuda_time_us": 2053.6769999999997, + "pct_cuda_time": 35.60876904150703, "invocations": 32 }, "children": [ { "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", - "cuda_time_us": 2051.6830000000004, - "pct_cuda_time": 35.74775400769498, + "cuda_time_us": 2053.6769999999997, + "pct_cuda_time": 35.60876904150703, "invocations": 32 }, "children": [] @@ -9987,16 +9987,16 @@ { "entry": { "name": "SiluAndMul", - "cuda_time_us": 272.49499999999995, - "pct_cuda_time": 4.747850534574219, + "cuda_time_us": 266.127, + "pct_cuda_time": 4.614384286676602, "invocations": 32 }, "children": [ { "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", - "cuda_time_us": 272.49499999999995, - "pct_cuda_time": 4.747850534574219, + "cuda_time_us": 266.127, + "pct_cuda_time": 4.614384286676602, "invocations": 32 }, "children": [] @@ -10006,16 +10006,16 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cuda_time_us": 1096.3029999999999, - "pct_cuda_time": 19.101571715463848, + "cuda_time_us": 1106.1129999999998, + "pct_cuda_time": 19.178927528919335, "invocations": 32 }, "children": [ { "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", - "cuda_time_us": 1096.3029999999999, - "pct_cuda_time": 19.101571715463848, + "cuda_time_us": 1106.1129999999998, + "pct_cuda_time": 19.178927528919335, "invocations": 32 }, "children": [] @@ -10029,16 +10029,16 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cuda_time_us": 4.326, - "pct_cuda_time": 0.07537459921307942, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.07917001526701674, "invocations": 1 }, "children": [ { "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", - "cuda_time_us": 4.326, - "pct_cuda_time": 0.07537459921307942, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.07917001526701674, "invocations": 1 }, "children": [] @@ -10050,16 +10050,16 @@ { "entry": { "name": "LogitsProcessor", - "cuda_time_us": 261.408, - "pct_cuda_time": 4.554674810701033, + "cuda_time_us": 265.255, + "pct_cuda_time": 4.599264651697881, "invocations": 1 }, "children": [ { "entry": { "name": "void at::native::(anonymous namespace)::indexSelectSmallIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, long)", - "cuda_time_us": 3.123, - "pct_cuda_time": 0.05441397904356151, + "cuda_time_us": 3.164, + "pct_cuda_time": 0.05486069389067915, "invocations": 1 }, "children": [] @@ -10067,8 +10067,8 @@ { "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", - "cuda_time_us": 258.285, - "pct_cuda_time": 4.500260831657472, + "cuda_time_us": 262.091, + "pct_cuda_time": 4.5444039578072015, "invocations": 1 }, "children": [] @@ -10078,16 +10078,16 @@ { "entry": { "name": "Sampler", - "cuda_time_us": 185.66100000000003, - "pct_cuda_time": 3.2348875322467734, + "cuda_time_us": 185.664, + "pct_cuda_time": 3.2192338402399026, "invocations": 1 }, "children": [ { "entry": { "name": "Memcpy DtoD (Device -> Device)", - "cuda_time_us": 19.099999999999998, - "pct_cuda_time": 0.3327912262990793, + "cuda_time_us": 17.14, + "pct_cuda_time": 0.29719099029274354, "invocations": 7 }, "children": [] @@ -10095,8 +10095,8 @@ { "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 3.164, - "pct_cuda_time": 0.05512834764451765, + "cuda_time_us": 3.524, + "pct_cuda_time": 0.06110274502868309, "invocations": 1 }, "children": [] @@ -10104,8 +10104,8 @@ { "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2}>(int, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2})", - "cuda_time_us": 9.333, - "pct_cuda_time": 0.1626146866517962, + "cuda_time_us": 10.254, + "pct_cuda_time": 0.17779442324747913, "invocations": 1 }, "children": [] @@ -10113,8 +10113,8 @@ { "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::SoftMaxForwardEpilogue>(float*, float const*, int)", - "cuda_time_us": 63.494, - "pct_cuda_time": 1.1062956085148556, + "cuda_time_us": 64.015, + "pct_cuda_time": 1.1099580655536743, "invocations": 1 }, "children": [] @@ -10122,8 +10122,8 @@ { "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::LogSoftMaxForwardEpilogue>(float*, float const*, int)", - "cuda_time_us": 35.973, - "pct_cuda_time": 0.6267800410291507, + "cuda_time_us": 35.733, + "pct_cuda_time": 0.6195755925397085, "invocations": 1 }, "children": [] @@ -10131,8 +10131,8 @@ { "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 3.804, - "pct_cuda_time": 0.06627946726919884, + "cuda_time_us": 3.844, + "pct_cuda_time": 0.06665123492913105, "invocations": 1 }, "children": [] @@ -10140,8 +10140,8 @@ { "entry": { "name": "void at::native::index_elementwise_kernel<128, 4, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1}>(long, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1})", - "cuda_time_us": 5.567, - "pct_cuda_time": 0.0969973171103128, + "cuda_time_us": 5.848, + "pct_cuda_time": 0.10139865293068634, "invocations": 1 }, "children": [] @@ -10149,8 +10149,8 @@ { "entry": { "name": "void at::native::reduce_kernel<512, 1, at::native::ReduceOp, unsigned int, long, 4, 4> >(at::native::ReduceOp, unsigned int, long, 4, 4>)", - "cuda_time_us": 40.78, - "pct_cuda_time": 0.7105354035851547, + "cuda_time_us": 40.58, + "pct_cuda_time": 0.7036178755005561, "invocations": 1 }, "children": [] @@ -10158,8 +10158,8 @@ { "entry": { "name": "Memcpy DtoH (Device -> Host)", - "cuda_time_us": 4.446, - "pct_cuda_time": 0.07746543414270715, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "invocations": 1 }, "children": [] @@ -10171,18 +10171,18 @@ { "entry": { "name": "LlamaForCausalLM", - "cpu_time_us": 17010.776, - "cuda_time_us": 5292.264999999999, - "pct_cuda_time": 92.2104376570522, + "cpu_time_us": 19988.312, + "cuda_time_us": 5316.416, + "pct_cuda_time": 92.18150150806221, "trace": "" }, "children": [ { "entry": { "name": "VocabParallelEmbedding(weight=bfloat16[128256, 4096])", - "cpu_time_us": 100.59, - "cuda_time_us": 3.844, - "pct_cuda_time": 0.0669764122457414, + "cpu_time_us": 133.19, + "cuda_time_us": 4.165, + "pct_cuda_time": 0.0722170638605179, "trace": "" }, "children": [ @@ -10190,8 +10190,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::indexSelectSmallIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, long)", "cpu_time_us": 0, - "cuda_time_us": 3.844, - "pct_cuda_time": 0.0669764122457414, + "cuda_time_us": 4.165, + "pct_cuda_time": 0.0722170638605179, "trace": "index_select(bfloat16[128256, 4096], 0, int64[1]) <- embedding(bfloat16[128256, 4096], int64[1], -1, False, False)" }, "children": [] @@ -10201,18 +10201,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 952.202, - "cuda_time_us": 185.667, - "pct_cuda_time": 3.2349920739932543, + "cpu_time_us": 1184.25, + "cuda_time_us": 187.791, + "pct_cuda_time": 3.256113959046943, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 119.438, - "cuda_time_us": 5.207, - "pct_cuda_time": 0.09072481232142963, + "cpu_time_us": 154.561, + "cuda_time_us": 5.206, + "pct_cuda_time": 0.09026699506791266, "trace": "" }, "children": [ @@ -10220,8 +10220,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::rms_norm_kernel(c10::BFloat16*, c10::BFloat16 const*, c10::BFloat16 const*, float, int, unsigned long, unsigned long)", "cpu_time_us": 0, - "cuda_time_us": 5.207, - "pct_cuda_time": 0.09072481232142963, + "cuda_time_us": 5.206, + "pct_cuda_time": 0.09026699506791266, "trace": "_C::rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10231,18 +10231,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 587.837, - "cuda_time_us": 54.797, - "pct_cuda_time": 0.9547623469900863, + "cpu_time_us": 749.02, + "cuda_time_us": 54.877, + "pct_cuda_time": 0.9515140008340074, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 145.587, - "cuda_time_us": 19.148, - "pct_cuda_time": 0.3336275602709304, + "cpu_time_us": 204.996, + "cuda_time_us": 18.467, + "pct_cuda_time": 0.32019988434866364, "trace": "" }, "children": [ @@ -10250,8 +10250,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 19.148, - "pct_cuda_time": 0.3336275602709304, + "cuda_time_us": 18.467, + "pct_cuda_time": 0.32019988434866364, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10261,9 +10261,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 51.567, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.0788593240957923, + "cpu_time_us": 65.278, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -10271,8 +10271,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.0788593240957923, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -10282,9 +10282,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 248.961, - "cuda_time_us": 17.103, - "pct_cuda_time": 0.29799624834519134, + "cpu_time_us": 286.468, + "cuda_time_us": 17.704, + "pct_cuda_time": 0.30697020374228307, "trace": "" }, "children": [ @@ -10292,8 +10292,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cuda_time_us": 3.405, + "pct_cuda_time": 0.059039400346954006, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10302,8 +10302,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 8.612, - "pct_cuda_time": 0.15005225344961629, + "cuda_time_us": 9.853, + "pct_cuda_time": 0.17084147184098028, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10312,8 +10312,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.07708933155434876, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10323,9 +10323,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 73.109, - "cuda_time_us": 14.02, - "pct_cuda_time": 0.24427921427817237, + "cpu_time_us": 78.988, + "cuda_time_us": 14.06, + "pct_cuda_time": 0.24378677500093196, "trace": "" }, "children": [ @@ -10333,8 +10333,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.02, - "pct_cuda_time": 0.24427921427817237, + "cuda_time_us": 14.06, + "pct_cuda_time": 0.24378677500093196, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10346,9 +10346,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 40.33, - "cuda_time_us": 5.287, - "pct_cuda_time": 0.09211870227451478, + "cpu_time_us": 46.6, + "cuda_time_us": 5.087, + "pct_cuda_time": 0.08820365038618357, "trace": "" }, "children": [ @@ -10356,8 +10356,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.287, - "pct_cuda_time": 0.09211870227451478, + "cuda_time_us": 5.087, + "pct_cuda_time": 0.08820365038618357, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10367,18 +10367,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 173.76, - "cuda_time_us": 120.376, - "pct_cuda_time": 2.097386212407224, + "cpu_time_us": 196.283, + "cuda_time_us": 122.62100000000001, + "pct_cuda_time": 2.1261293127588394, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 59.489, - "cuda_time_us": 73.268, - "pct_cuda_time": 1.2765941135330336, + "cpu_time_us": 68.843, + "cuda_time_us": 75.351, + "pct_cuda_time": 1.306513320277043, "trace": "" }, "children": [ @@ -10386,8 +10386,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 73.268, - "pct_cuda_time": 1.2765941135330336, + "cuda_time_us": 75.351, + "pct_cuda_time": 1.306513320277043, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10397,9 +10397,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 29.785, - "cuda_time_us": 9.253, - "pct_cuda_time": 0.16122079669871103, + "cpu_time_us": 35.934, + "cuda_time_us": 9.414, + "pct_cuda_time": 0.16322963725880324, "trace": "" }, "children": [ @@ -10407,8 +10407,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 9.253, - "pct_cuda_time": 0.16122079669871103, + "cuda_time_us": 9.414, + "pct_cuda_time": 0.16322963725880324, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -10418,9 +10418,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 57.997, - "cuda_time_us": 37.855, - "pct_cuda_time": 0.6595713021754789, + "cpu_time_us": 66.56, + "cuda_time_us": 37.856, + "pct_cuda_time": 0.6563863552229929, "trace": "" }, "children": [ @@ -10428,8 +10428,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 37.855, - "pct_cuda_time": 0.6595713021754789, + "cuda_time_us": 37.856, + "pct_cuda_time": 0.6563863552229929, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -10443,18 +10443,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 558.274, - "cuda_time_us": 181.379, - "pct_cuda_time": 3.16027957250789, + "cpu_time_us": 635.38, + "cuda_time_us": 184.826, + "pct_cuda_time": 3.2047037323131042, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 29.334, - "cuda_time_us": 5.006, - "pct_cuda_time": 0.0872226638143032, + "cpu_time_us": 31.117, + "cuda_time_us": 4.846, + "pct_cuda_time": 0.08402494392990871, "trace": "" }, "children": [ @@ -10462,8 +10462,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.006, - "pct_cuda_time": 0.0872226638143032, + "cuda_time_us": 4.846, + "pct_cuda_time": 0.08402494392990871, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10473,18 +10473,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 328.941, - "cuda_time_us": 52.191, - "pct_cuda_time": 0.9093563817683377, + "cpu_time_us": 382.843, + "cuda_time_us": 54.635999999999996, + "pct_cuda_time": 0.9473352943777325, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 52.749, - "cuda_time_us": 19.468, - "pct_cuda_time": 0.339203120083271, + "cpu_time_us": 56.465, + "cuda_time_us": 18.627, + "pct_cuda_time": 0.3229741292988876, "trace": "" }, "children": [ @@ -10492,8 +10492,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 19.468, - "pct_cuda_time": 0.339203120083271, + "cuda_time_us": 18.627, + "pct_cuda_time": 0.3229741292988876, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10503,9 +10503,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 25.168, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 31.948, + "cuda_time_us": 4.406, + "pct_cuda_time": 0.07639577031679276, "trace": "" }, "children": [ @@ -10513,8 +10513,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.406, + "pct_cuda_time": 0.07639577031679276, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -10524,9 +10524,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 150.595, - "cuda_time_us": 14.178, - "pct_cuda_time": 0.2470321469355155, + "cpu_time_us": 186.239, + "cuda_time_us": 16.702, + "pct_cuda_time": 0.28959649474150545, "trace": "" }, "children": [ @@ -10534,8 +10534,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.163, - "pct_cuda_time": 0.05511092402010408, + "cuda_time_us": 3.204, + "pct_cuda_time": 0.05555425512823514, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10544,8 +10544,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.769, - "pct_cuda_time": 0.11794051365541718, + "cuda_time_us": 9.333, + "pct_cuda_time": 0.16182517575275235, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10554,8 +10554,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.246, - "pct_cuda_time": 0.0739807092599943, + "cuda_time_us": 4.165, + "pct_cuda_time": 0.0722170638605179, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10565,9 +10565,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 68.041, - "cuda_time_us": 13.979, - "pct_cuda_time": 0.2435648456772162, + "cpu_time_us": 67.07, + "cuda_time_us": 14.901, + "pct_cuda_time": 0.25836890002054674, "trace": "" }, "children": [ @@ -10575,8 +10575,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 13.979, - "pct_cuda_time": 0.2435648456772162, + "cuda_time_us": 14.901, + "pct_cuda_time": 0.25836890002054674, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10588,9 +10588,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 30.285, - "cuda_time_us": 5.007, - "pct_cuda_time": 0.08724008743871675, + "cpu_time_us": 38.338, + "cuda_time_us": 5.047, + "pct_cuda_time": 0.08751008914862757, "trace": "" }, "children": [ @@ -10598,8 +10598,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.007, - "pct_cuda_time": 0.08724008743871675, + "cuda_time_us": 5.047, + "pct_cuda_time": 0.08751008914862757, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10609,18 +10609,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 146.809, - "cuda_time_us": 119.17499999999998, - "pct_cuda_time": 2.0764604394865325, + "cpu_time_us": 154.952, + "cuda_time_us": 120.297, + "pct_cuda_time": 2.085833404856836, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 49.975, - "cuda_time_us": 73.509, - "pct_cuda_time": 1.2807932070167025, + "cpu_time_us": 52.97, + "cuda_time_us": 74.27, + "pct_cuda_time": 1.2877698278320922, "trace": "" }, "children": [ @@ -10628,8 +10628,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 73.509, - "pct_cuda_time": 1.2807932070167025, + "cuda_time_us": 74.27, + "pct_cuda_time": 1.2877698278320922, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10639,9 +10639,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 28.553, - "cuda_time_us": 8.692, - "pct_cuda_time": 0.15144614340270143, + "cpu_time_us": 23.074, + "cuda_time_us": 8.732, + "pct_cuda_time": 0.15140441815847352, "trace": "" }, "children": [ @@ -10649,8 +10649,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.692, - "pct_cuda_time": 0.15144614340270143, + "cuda_time_us": 8.732, + "pct_cuda_time": 0.15140441815847352, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -10660,9 +10660,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 52.027, - "cuda_time_us": 36.974, - "pct_cuda_time": 0.6442210890671286, + "cpu_time_us": 59.629, + "cuda_time_us": 37.295, + "pct_cuda_time": 0.6466591588662701, "trace": "" }, "children": [ @@ -10670,8 +10670,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 36.974, - "pct_cuda_time": 0.6442210890671286, + "cuda_time_us": 37.295, + "pct_cuda_time": 0.6466591588662701, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -10685,18 +10685,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 524.864, - "cuda_time_us": 182.662, - "pct_cuda_time": 3.1826340826304937, + "cpu_time_us": 632.195, + "cuda_time_us": 182.462, + "pct_cuda_time": 3.1637142631735458, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 31.558, - "cuda_time_us": 4.966, - "pct_cuda_time": 0.08652571883776063, + "cpu_time_us": 30.035, + "cuda_time_us": 5.007, + "pct_cuda_time": 0.08681652791107157, "trace": "" }, "children": [ @@ -10704,8 +10704,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.966, - "pct_cuda_time": 0.08652571883776063, + "cuda_time_us": 5.007, + "pct_cuda_time": 0.08681652791107157, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10715,18 +10715,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 308.4, - "cuda_time_us": 53.033, - "pct_cuda_time": 0.9240270735245588, + "cpu_time_us": 370.223, + "cuda_time_us": 51.87100000000001, + "pct_cuda_time": 0.8993928738316747, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 50.335, - "cuda_time_us": 20.309, - "pct_cuda_time": 0.35385638821507864, + "cpu_time_us": 56.264, + "cuda_time_us": 18.346, + "pct_cuda_time": 0.31810186160505677, "trace": "" }, "children": [ @@ -10734,8 +10734,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 20.309, - "pct_cuda_time": 0.35385638821507864, + "cuda_time_us": 18.346, + "pct_cuda_time": 0.31810186160505677, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10745,9 +10745,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 24.967, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.0788593240957923, + "cpu_time_us": 31.036, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.07708933155434876, "trace": "" }, "children": [ @@ -10755,8 +10755,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.0788593240957923, + "cuda_time_us": 4.446, + "pct_cuda_time": 0.07708933155434876, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -10766,9 +10766,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 142.643, - "cuda_time_us": 14.218, - "pct_cuda_time": 0.24772909191205808, + "cpu_time_us": 179.559, + "cuda_time_us": 15.099, + "pct_cuda_time": 0.26180202814644893, "trace": "" }, "children": [ @@ -10776,8 +10776,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.404, - "pct_cuda_time": 0.059310017503773085, + "cuda_time_us": 3.284, + "pct_cuda_time": 0.056941377603347124, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10786,8 +10786,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.729, - "pct_cuda_time": 0.1172435686788746, + "cuda_time_us": 7.81, + "pct_cuda_time": 0.13541783163280788, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10796,8 +10796,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.085, - "pct_cuda_time": 0.07117550572941042, + "cuda_time_us": 4.005, + "pct_cuda_time": 0.06944281891029393, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -10807,9 +10807,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 59.359, + "cpu_time_us": 64.427, "cuda_time_us": 13.98, - "pct_cuda_time": 0.2435822693016298, + "pct_cuda_time": 0.24239965252582002, "trace": "" }, "children": [ @@ -10818,7 +10818,7 @@ "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, "cuda_time_us": 13.98, - "pct_cuda_time": 0.2435822693016298, + "pct_cuda_time": 0.24239965252582002, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10830,9 +10830,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 27.431, - "cuda_time_us": 5.167, - "pct_cuda_time": 0.09002786734488706, + "cpu_time_us": 31.918, + "cuda_time_us": 5.287, + "pct_cuda_time": 0.09167145657396354, "trace": "" }, "children": [ @@ -10840,8 +10840,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.167, - "pct_cuda_time": 0.09002786734488706, + "cuda_time_us": 5.287, + "pct_cuda_time": 0.09167145657396354, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10851,18 +10851,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 137.205, - "cuda_time_us": 119.496, - "pct_cuda_time": 2.082053422923287, + "cpu_time_us": 171.476, + "cuda_time_us": 120.297, + "pct_cuda_time": 2.085833404856836, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.912, - "cuda_time_us": 73.669, - "pct_cuda_time": 1.283580986922873, + "cpu_time_us": 61.792, + "cuda_time_us": 74.511, + "pct_cuda_time": 1.291948534288367, "trace": "" }, "children": [ @@ -10870,8 +10870,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 73.669, - "pct_cuda_time": 1.283580986922873, + "cuda_time_us": 74.511, + "pct_cuda_time": 1.291948534288367, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10881,9 +10881,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.927, - "cuda_time_us": 8.812, - "pct_cuda_time": 0.15353697833232913, + "cpu_time_us": 22.814, + "cuda_time_us": 8.892, + "pct_cuda_time": 0.15417866310869752, "trace": "" }, "children": [ @@ -10891,8 +10891,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.812, - "pct_cuda_time": 0.15353697833232913, + "cuda_time_us": 8.892, + "pct_cuda_time": 0.15417866310869752, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -10902,9 +10902,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 50.876, - "cuda_time_us": 37.015, - "pct_cuda_time": 0.6449354576680849, + "cpu_time_us": 67.551, + "cuda_time_us": 36.894, + "pct_cuda_time": 0.6397062074597712, "trace": "" }, "children": [ @@ -10912,8 +10912,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 37.015, - "pct_cuda_time": 0.6449354576680849, + "cuda_time_us": 36.894, + "pct_cuda_time": 0.6397062074597712, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -10927,18 +10927,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 484.804, - "cuda_time_us": 181.019, - "pct_cuda_time": 3.1540070677190073, + "cpu_time_us": 571.555, + "cuda_time_us": 161.308, + "pct_cuda_time": 2.7969244026920577, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.147, - "cuda_time_us": 5.367, - "pct_cuda_time": 0.09351259222759993, + "cpu_time_us": 33.821, + "cuda_time_us": 4.926, + "pct_cuda_time": 0.08541206640502069, "trace": "" }, "children": [ @@ -10946,8 +10946,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.367, - "pct_cuda_time": 0.09351259222759993, + "cuda_time_us": 4.926, + "pct_cuda_time": 0.08541206640502069, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -10957,18 +10957,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 291.506, - "cuda_time_us": 51.351, - "pct_cuda_time": 0.8947205372609436, + "cpu_time_us": 341.13, + "cuda_time_us": 48.584999999999994, + "pct_cuda_time": 0.8424168181664493, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 47.691, - "cuda_time_us": 18.787, - "pct_cuda_time": 0.32733763185763365, + "cpu_time_us": 56.435, + "cuda_time_us": 17.906, + "pct_cuda_time": 0.31047268799194083, "trace": "" }, "children": [ @@ -10976,8 +10976,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 18.787, - "pct_cuda_time": 0.32733763185763365, + "cuda_time_us": 17.906, + "pct_cuda_time": 0.31047268799194083, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -10987,9 +10987,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.704, - "cuda_time_us": 4.165, - "pct_cuda_time": 0.07256939568249557, + "cpu_time_us": 27.842, + "cuda_time_us": 4.205, + "pct_cuda_time": 0.0729106250980739, "trace": "" }, "children": [ @@ -10997,8 +10997,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.165, - "pct_cuda_time": 0.07256939568249557, + "cuda_time_us": 4.205, + "pct_cuda_time": 0.0729106250980739, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -11008,9 +11008,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 137.035, - "cuda_time_us": 14.578999999999999, - "pct_cuda_time": 0.2540190203253548, + "cpu_time_us": 163.494, + "cuda_time_us": 14.738, + "pct_cuda_time": 0.2555426379775061, "trace": "" }, "children": [ @@ -11018,8 +11018,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.284, - "pct_cuda_time": 0.057219182574145364, + "cuda_time_us": 2.963, + "pct_cuda_time": 0.05137554867196027, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11028,8 +11028,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 7.25, - "pct_cuda_time": 0.12632127699834164, + "cuda_time_us": 7.85, + "pct_cuda_time": 0.13611139287036386, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11038,8 +11038,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11049,9 +11049,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.894, - "cuda_time_us": 13.82, - "pct_cuda_time": 0.24079448939545947, + "cpu_time_us": 58.758, + "cuda_time_us": 11.736, + "pct_cuda_time": 0.20349086709892875, "trace": "" }, "children": [ @@ -11059,8 +11059,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 13.82, - "pct_cuda_time": 0.24079448939545947, + "cuda_time_us": 11.736, + "pct_cuda_time": 0.20349086709892875, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11072,9 +11072,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.328, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cpu_time_us": 29.945, + "cuda_time_us": 4.806, + "pct_cuda_time": 0.08333138269235271, "trace": "" }, "children": [ @@ -11082,8 +11082,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cuda_time_us": 4.806, + "pct_cuda_time": 0.08333138269235271, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11093,18 +11093,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 126.298, - "cuda_time_us": 119.41499999999999, - "pct_cuda_time": 2.080642109345788, + "cpu_time_us": 146.379, + "cuda_time_us": 102.991, + "pct_cuda_time": 1.785764135428235, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 46.61, - "cuda_time_us": 73.429, - "pct_cuda_time": 1.2793993170636175, + "cpu_time_us": 51.727, + "cuda_time_us": 62.653, + "pct_cuda_time": 1.0863423054148926, "trace": "" }, "children": [ @@ -11112,8 +11112,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 73.429, - "pct_cuda_time": 1.2793993170636175, + "cuda_time_us": 62.653, + "pct_cuda_time": 1.0863423054148926, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11123,9 +11123,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.947, - "cuda_time_us": 8.612, - "pct_cuda_time": 0.15005225344961629, + "cpu_time_us": 19.94, + "cuda_time_us": 8.451, + "pct_cuda_time": 0.14653215046464269, "trace": "" }, "children": [ @@ -11133,8 +11133,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.612, - "pct_cuda_time": 0.15005225344961629, + "cuda_time_us": 8.451, + "pct_cuda_time": 0.14653215046464269, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -11144,9 +11144,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 47.0, - "cuda_time_us": 37.374, - "pct_cuda_time": 0.6511905388325545, + "cpu_time_us": 58.247, + "cuda_time_us": 31.887, + "pct_cuda_time": 0.5528896795486997, "trace": "" }, "children": [ @@ -11154,8 +11154,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 37.374, - "pct_cuda_time": 0.6511905388325545, + "cuda_time_us": 31.887, + "pct_cuda_time": 0.5528896795486997, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -11169,18 +11169,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 501.038, - "cuda_time_us": 161.912, - "pct_cuda_time": 2.821093876049033, + "cpu_time_us": 639.526, + "cuda_time_us": 161.249, + "pct_cuda_time": 2.7959013998666626, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.376, - "cuda_time_us": 4.486, - "pct_cuda_time": 0.07816237911924973, + "cpu_time_us": 27.711, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -11188,8 +11188,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.486, - "pct_cuda_time": 0.07816237911924973, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11199,18 +11199,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 301.811, - "cuda_time_us": 49.989000000000004, - "pct_cuda_time": 0.8709895608096689, + "cpu_time_us": 390.304, + "cuda_time_us": 46.903000000000006, + "pct_cuda_time": 0.81325256812722, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 50.655, - "cuda_time_us": 19.308, - "pct_cuda_time": 0.33641534017710073, + "cpu_time_us": 74.562, + "cuda_time_us": 16.103, + "pct_cuda_time": 0.2792104152091044, "trace": "" }, "children": [ @@ -11218,8 +11218,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 19.308, - "pct_cuda_time": 0.33641534017710073, + "cuda_time_us": 16.103, + "pct_cuda_time": 0.2792104152091044, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11229,9 +11229,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.873, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cpu_time_us": 28.042, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "" }, "children": [ @@ -11239,8 +11239,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -11250,9 +11250,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 138.146, - "cuda_time_us": 14.498999999999999, - "pct_cuda_time": 0.25262513037226964, + "cpu_time_us": 174.331, + "cuda_time_us": 14.939, + "pct_cuda_time": 0.25902778319622494, "trace": "" }, "children": [ @@ -11261,7 +11261,7 @@ "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, "cuda_time_us": 2.803, - "pct_cuda_time": 0.048838419231220905, + "pct_cuda_time": 0.048601303721736294, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11270,8 +11270,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 7.49, - "pct_cuda_time": 0.13050294685759706, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11280,8 +11280,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.206, - "pct_cuda_time": 0.07328376428345171, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11291,9 +11291,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 61.572, - "cuda_time_us": 12.137, - "pct_cuda_time": 0.21147052950743067, + "cpu_time_us": 64.606, + "cuda_time_us": 12.057, + "pct_cuda_time": 0.2090566960303156, "trace": "" }, "children": [ @@ -11301,8 +11301,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 12.137, - "pct_cuda_time": 0.21147052950743067, + "cuda_time_us": 12.057, + "pct_cuda_time": 0.2090566960303156, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11314,9 +11314,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.36, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.0844348839081329, + "cpu_time_us": 31.297, + "cuda_time_us": 4.926, + "pct_cuda_time": 0.08541206640502069, "trace": "" }, "children": [ @@ -11324,8 +11324,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.846, - "pct_cuda_time": 0.0844348839081329, + "cuda_time_us": 4.926, + "pct_cuda_time": 0.08541206640502069, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11335,18 +11335,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 130.605, - "cuda_time_us": 102.59100000000001, - "pct_cuda_time": 1.7875070522119816, + "cpu_time_us": 168.492, + "cuda_time_us": 104.654, + "pct_cuda_time": 1.8145989438796255, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 48.502, - "cuda_time_us": 63.214, - "pct_cuda_time": 1.1014169936790574, + "cpu_time_us": 69.664, + "cuda_time_us": 62.813, + "pct_cuda_time": 1.0891165503651168, "trace": "" }, "children": [ @@ -11354,8 +11354,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.214, - "pct_cuda_time": 1.1014169936790574, + "cuda_time_us": 62.813, + "pct_cuda_time": 1.0891165503651168, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11365,9 +11365,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.366, - "cuda_time_us": 8.572, - "pct_cuda_time": 0.14935530847307368, + "cpu_time_us": 20.61, + "cuda_time_us": 8.292, + "pct_cuda_time": 0.1437752445453576, "trace": "" }, "children": [ @@ -11375,8 +11375,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.572, - "pct_cuda_time": 0.14935530847307368, + "cuda_time_us": 8.292, + "pct_cuda_time": 0.1437752445453576, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -11386,9 +11386,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 49.454, - "cuda_time_us": 30.805, - "pct_cuda_time": 0.5367347500598502, + "cpu_time_us": 60.24, + "cuda_time_us": 33.549, + "pct_cuda_time": 0.5817071489691512, "trace": "" }, "children": [ @@ -11396,8 +11396,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 30.805, - "pct_cuda_time": 0.5367347500598502, + "cuda_time_us": 33.549, + "pct_cuda_time": 0.5817071489691512, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -11411,18 +11411,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 515.599, - "cuda_time_us": 157.022, - "pct_cuda_time": 2.735892352666703, + "cpu_time_us": 580.868, + "cuda_time_us": 167.27, + "pct_cuda_time": 2.900299705149779, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.627, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cpu_time_us": 28.933, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -11430,8 +11430,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11441,18 +11441,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 297.354, - "cuda_time_us": 46.022, - "pct_cuda_time": 0.8018700427610591, + "cpu_time_us": 327.37, + "cuda_time_us": 51.492000000000004, + "pct_cuda_time": 0.8928213811058315, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 48.973, - "cuda_time_us": 16.544, - "pct_cuda_time": 0.28825644229800884, + "cpu_time_us": 59.088, + "cuda_time_us": 18.247, + "pct_cuda_time": 0.3163852975421057, "trace": "" }, "children": [ @@ -11460,8 +11460,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.544, - "pct_cuda_time": 0.28825644229800884, + "cuda_time_us": 18.247, + "pct_cuda_time": 0.3163852975421057, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11471,9 +11471,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 23.586, - "cuda_time_us": 4.005, - "pct_cuda_time": 0.06978161577632526, + "cpu_time_us": 24.266, + "cuda_time_us": 3.724, + "pct_cuda_time": 0.06457055121646307, "trace": "" }, "children": [ @@ -11481,8 +11481,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.005, - "pct_cuda_time": 0.06978161577632526, + "cuda_time_us": 3.724, + "pct_cuda_time": 0.06457055121646307, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -11492,9 +11492,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 141.591, - "cuda_time_us": 13.016000000000002, - "pct_cuda_time": 0.22678589536695376, + "cpu_time_us": 152.939, + "cuda_time_us": 15.661000000000001, + "pct_cuda_time": 0.27154656353411066, "trace": "" }, "children": [ @@ -11502,8 +11502,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.883, - "pct_cuda_time": 0.05023230918430605, + "cuda_time_us": 2.643, + "pct_cuda_time": 0.045827058771512316, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11512,8 +11512,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.048, - "pct_cuda_time": 0.10537808045323727, + "cuda_time_us": 9.253, + "pct_cuda_time": 0.16043805327764035, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11522,8 +11522,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.085, - "pct_cuda_time": 0.07117550572941042, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11533,9 +11533,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.753, - "cuda_time_us": 12.457, - "pct_cuda_time": 0.21704608931977126, + "cpu_time_us": 59.018, + "cuda_time_us": 13.86, + "pct_cuda_time": 0.240318968813152, "trace": "" }, "children": [ @@ -11543,8 +11543,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 12.457, - "pct_cuda_time": 0.21704608931977126, + "cuda_time_us": 13.86, + "pct_cuda_time": 0.240318968813152, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11556,9 +11556,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 30.135, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "cpu_time_us": 30.796, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -11566,8 +11566,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11577,18 +11577,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 141.712, - "cuda_time_us": 101.388, - "pct_cuda_time": 1.7665464320424638, + "cpu_time_us": 158.186, + "cuda_time_us": 106.366, + "pct_cuda_time": 1.844283364847022, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 53.409, - "cuda_time_us": 61.211, - "pct_cuda_time": 1.066517473978688, + "cpu_time_us": 57.656, + "cuda_time_us": 64.124, + "pct_cuda_time": 1.1118480199260143, "trace": "" }, "children": [ @@ -11596,8 +11596,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.211, - "pct_cuda_time": 1.066517473978688, + "cuda_time_us": 64.124, + "pct_cuda_time": 1.1118480199260143, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11607,9 +11607,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 18.357, - "cuda_time_us": 8.451, - "pct_cuda_time": 0.14724704991903242, + "cpu_time_us": 19.509, + "cuda_time_us": 8.132, + "pct_cuda_time": 0.14100099959513362, "trace": "" }, "children": [ @@ -11617,8 +11617,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.451, - "pct_cuda_time": 0.14724704991903242, + "cuda_time_us": 8.132, + "pct_cuda_time": 0.14100099959513362, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -11628,9 +11628,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 50.805, - "cuda_time_us": 31.726, - "pct_cuda_time": 0.5527819081447429, + "cpu_time_us": 62.243, + "cuda_time_us": 34.11, + "pct_cuda_time": 0.591434345325874, "trace": "" }, "children": [ @@ -11638,8 +11638,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 31.726, - "pct_cuda_time": 0.5527819081447429, + "cuda_time_us": 34.11, + "pct_cuda_time": 0.591434345325874, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -11653,18 +11653,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 566.066, - "cuda_time_us": 165.565, - "pct_cuda_time": 2.884742376031783, + "cpu_time_us": 633.187, + "cuda_time_us": 162.742, + "pct_cuda_time": 2.82178857305844, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.747, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 30.345, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -11672,8 +11672,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11683,18 +11683,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 369.712, - "cuda_time_us": 50.348, - "pct_cuda_time": 0.8772446419741385, + "cpu_time_us": 414.079, + "cuda_time_us": 47.024, + "pct_cuda_time": 0.8153505908708268, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 89.253, - "cuda_time_us": 18.336, - "pct_cuda_time": 0.3194795772471161, + "cpu_time_us": 116.855, + "cuda_time_us": 17.365, + "pct_cuda_time": 0.301092272253996, "trace": "" }, "children": [ @@ -11702,8 +11702,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 18.336, - "pct_cuda_time": 0.3194795772471161, + "cuda_time_us": 17.365, + "pct_cuda_time": 0.301092272253996, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11713,9 +11713,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 24.236, - "cuda_time_us": 4.205, - "pct_cuda_time": 0.07326634065903814, + "cpu_time_us": 27.311, + "cuda_time_us": 4.045, + "pct_cuda_time": 0.07013638014784991, "trace": "" }, "children": [ @@ -11723,8 +11723,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.205, - "pct_cuda_time": 0.07326634065903814, + "cuda_time_us": 4.045, + "pct_cuda_time": 0.07013638014784991, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -11734,9 +11734,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 168.952, - "cuda_time_us": 13.216000000000001, - "pct_cuda_time": 0.23027062024966663, + "cpu_time_us": 174.341, + "cuda_time_us": 12.015, + "pct_cuda_time": 0.2083284567308818, "trace": "" }, "children": [ @@ -11744,8 +11744,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.843, - "pct_cuda_time": 0.04953536420776348, + "cuda_time_us": 2.923, + "pct_cuda_time": 0.050681987434404274, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11754,8 +11754,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.408, - "pct_cuda_time": 0.11165058524212045, + "cuda_time_us": 5.968, + "pct_cuda_time": 0.10347933664335435, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11764,8 +11764,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.124, + "pct_cuda_time": 0.05416713265312315, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11775,9 +11775,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 57.085, - "cuda_time_us": 14.591, - "pct_cuda_time": 0.25422810381831756, + "cpu_time_us": 59.929, + "cuda_time_us": 13.599, + "pct_cuda_time": 0.23579348173809914, "trace": "" }, "children": [ @@ -11785,8 +11785,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.591, - "pct_cuda_time": 0.25422810381831756, + "cuda_time_us": 13.599, + "pct_cuda_time": 0.23579348173809914, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11798,9 +11798,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.049, + "cpu_time_us": 27.19, "cuda_time_us": 4.766, - "pct_cuda_time": 0.08304099395504776, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -11809,7 +11809,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 4.766, - "pct_cuda_time": 0.08304099395504776, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11819,18 +11819,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 129.143, - "cuda_time_us": 105.76500000000001, - "pct_cuda_time": 1.8428096361006352, + "cpu_time_us": 142.103, + "cuda_time_us": 106.226, + "pct_cuda_time": 1.8418559005155761, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.43, - "cuda_time_us": 63.264, - "pct_cuda_time": 1.102288174899736, + "cpu_time_us": 52.749, + "cuda_time_us": 62.843, + "pct_cuda_time": 1.0896367212932836, "trace": "" }, "children": [ @@ -11838,8 +11838,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.264, - "pct_cuda_time": 1.102288174899736, + "cuda_time_us": 62.843, + "pct_cuda_time": 1.0896367212932836, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11849,9 +11849,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.847, - "cuda_time_us": 8.211, - "pct_cuda_time": 0.14306538005977698, + "cpu_time_us": 19.449, + "cuda_time_us": 8.532, + "pct_cuda_time": 0.14793661197069358, "trace": "" }, "children": [ @@ -11859,8 +11859,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.211, - "pct_cuda_time": 0.14306538005977698, + "cuda_time_us": 8.532, + "pct_cuda_time": 0.14793661197069358, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -11870,9 +11870,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 49.184, - "cuda_time_us": 34.29, - "pct_cuda_time": 0.597456081141122, + "cpu_time_us": 53.249, + "cuda_time_us": 34.851, + "pct_cuda_time": 0.6042825672515988, "trace": "" }, "children": [ @@ -11880,8 +11880,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 34.29, - "pct_cuda_time": 0.597456081141122, + "cuda_time_us": 34.851, + "pct_cuda_time": 0.6042825672515988, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -11895,18 +11895,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 487.197, - "cuda_time_us": 159.497, - "pct_cuda_time": 2.7790158230902753, + "cpu_time_us": 575.86, + "cuda_time_us": 162.78400000000002, + "pct_cuda_time": 2.8225168123578745, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.835, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cpu_time_us": 27.131, + "cuda_time_us": 4.807, + "pct_cuda_time": 0.08334872172329162, "trace": "" }, "children": [ @@ -11914,8 +11914,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cuda_time_us": 4.807, + "pct_cuda_time": 0.08334872172329162, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -11925,18 +11925,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 293.828, - "cuda_time_us": 48.566, - "pct_cuda_time": 0.8461957432691668, + "cpu_time_us": 351.496, + "cuda_time_us": 47.005, + "pct_cuda_time": 0.8150211492829876, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 47.891, - "cuda_time_us": 17.425, - "pct_cuda_time": 0.303606655406359, + "cpu_time_us": 58.076, + "cuda_time_us": 16.764, + "pct_cuda_time": 0.2906715146597172, "trace": "" }, "children": [ @@ -11944,8 +11944,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.425, - "pct_cuda_time": 0.303606655406359, + "cuda_time_us": 16.764, + "pct_cuda_time": 0.2906715146597172, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -11955,9 +11955,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 23.715, + "cpu_time_us": 26.449, "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "pct_cuda_time": 0.06736213519762593, "trace": "" }, "children": [ @@ -11966,7 +11966,7 @@ "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "pct_cuda_time": 0.06736213519762593, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -11976,9 +11976,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 135.102, - "cuda_time_us": 12.735999999999999, - "pct_cuda_time": 0.2219072805311557, + "cpu_time_us": 173.289, + "cuda_time_us": 13.978, + "pct_cuda_time": 0.24236497446394217, "trace": "" }, "children": [ @@ -11986,8 +11986,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.803, - "pct_cuda_time": 0.048838419231220905, + "cuda_time_us": 2.884, + "pct_cuda_time": 0.050005765227787186, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -11996,8 +11996,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.048, - "pct_cuda_time": 0.10537808045323727, + "cuda_time_us": 7.33, + "pct_cuda_time": 0.12709509678213593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12006,8 +12006,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12017,9 +12017,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 54.371, - "cuda_time_us": 14.52, - "pct_cuda_time": 0.25299102648495453, + "cpu_time_us": 61.672, + "cuda_time_us": 12.378, + "pct_cuda_time": 0.2146225249617024, "trace": "" }, "children": [ @@ -12027,8 +12027,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.52, - "pct_cuda_time": 0.25299102648495453, + "cuda_time_us": 12.378, + "pct_cuda_time": 0.2146225249617024, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12040,9 +12040,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.038, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 28.172, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -12050,8 +12050,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12061,18 +12061,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 126.649, - "cuda_time_us": 101.43900000000001, - "pct_cuda_time": 1.7674350368875553, + "cpu_time_us": 145.027, + "cuda_time_us": 106.24600000000001, + "pct_cuda_time": 1.8422026811343541, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 48.122, - "cuda_time_us": 61.481, - "pct_cuda_time": 1.0712218525703505, + "cpu_time_us": 53.289, + "cuda_time_us": 63.033, + "pct_cuda_time": 1.0929311371716746, "trace": "" }, "children": [ @@ -12080,8 +12080,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.481, - "pct_cuda_time": 1.0712218525703505, + "cuda_time_us": 63.033, + "pct_cuda_time": 1.0929311371716746, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12091,9 +12091,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.396, - "cuda_time_us": 8.091, - "pct_cuda_time": 0.14097454513014926, + "cpu_time_us": 19.389, + "cuda_time_us": 8.292, + "pct_cuda_time": 0.1437752445453576, "trace": "" }, "children": [ @@ -12101,8 +12101,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.091, - "pct_cuda_time": 0.14097454513014926, + "cuda_time_us": 8.292, + "pct_cuda_time": 0.1437752445453576, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -12112,9 +12112,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.689, - "cuda_time_us": 31.867, - "pct_cuda_time": 0.5552386391870555, + "cpu_time_us": 55.173, + "cuda_time_us": 34.921, + "pct_cuda_time": 0.6054962994173219, "trace": "" }, "children": [ @@ -12122,8 +12122,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 31.867, - "pct_cuda_time": 0.5552386391870555, + "cuda_time_us": 34.921, + "pct_cuda_time": 0.6054962994173219, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -12137,18 +12137,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 472.526, - "cuda_time_us": 164.274, - "pct_cuda_time": 2.862248476913872, + "cpu_time_us": 586.407, + "cuda_time_us": 166.809, + "pct_cuda_time": 2.892306411886946, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 22.944, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.511, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -12156,8 +12156,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12167,18 +12167,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 281.971, - "cuda_time_us": 49.408, - "pct_cuda_time": 0.860866435025388, + "cpu_time_us": 342.351, + "cuda_time_us": 49.609, + "pct_cuda_time": 0.860171985847883, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 45.208, - "cuda_time_us": 17.585, - "pct_cuda_time": 0.3063944353125293, + "cpu_time_us": 53.12, + "cuda_time_us": 17.666, + "pct_cuda_time": 0.30631132056660487, "trace": "" }, "children": [ @@ -12186,8 +12186,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.585, - "pct_cuda_time": 0.3063944353125293, + "cuda_time_us": 17.666, + "pct_cuda_time": 0.30631132056660487, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12197,9 +12197,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.832, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cpu_time_us": 25.568, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "" }, "children": [ @@ -12207,8 +12207,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -12218,9 +12218,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 132.999, - "cuda_time_us": 13.136999999999999, - "pct_cuda_time": 0.22889415392099502, + "cpu_time_us": 170.345, + "cuda_time_us": 14.699000000000002, + "pct_cuda_time": 0.25486641577088903, "trace": "" }, "children": [ @@ -12228,8 +12228,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.763, - "pct_cuda_time": 0.04814147425467833, + "cuda_time_us": 2.803, + "pct_cuda_time": 0.048601303721736294, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12238,8 +12238,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.329, - "pct_cuda_time": 0.11027411891344885, + "cuda_time_us": 7.811, + "pct_cuda_time": 0.13543517066374677, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12248,8 +12248,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 4.085, + "pct_cuda_time": 0.07082994138540591, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12259,9 +12259,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 53.7, - "cuda_time_us": 14.801, - "pct_cuda_time": 0.2578870649451661, + "cpu_time_us": 61.071, + "cuda_time_us": 13.479, + "pct_cuda_time": 0.23371279802543113, "trace": "" }, "children": [ @@ -12269,8 +12269,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.801, - "pct_cuda_time": 0.2578870649451661, + "cuda_time_us": 13.479, + "pct_cuda_time": 0.23371279802543113, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12282,9 +12282,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.738, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 28.532, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.07917001526701674, "trace": "" }, "children": [ @@ -12292,8 +12292,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.07917001526701674, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12303,18 +12303,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 125.337, - "cuda_time_us": 105.614, - "pct_cuda_time": 1.8401786688141866, + "cpu_time_us": 166.659, + "cuda_time_us": 107.868, + "pct_cuda_time": 1.8703265893172494, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.761, - "cuda_time_us": 63.173, - "pct_cuda_time": 1.1007026250781016, + "cpu_time_us": 55.132, + "cuda_time_us": 64.255, + "pct_cuda_time": 1.11411943297901, "trace": "" }, "children": [ @@ -12322,8 +12322,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.173, - "pct_cuda_time": 1.1007026250781016, + "cuda_time_us": 64.255, + "pct_cuda_time": 1.11411943297901, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12333,9 +12333,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.346, - "cuda_time_us": 8.452, - "pct_cuda_time": 0.147264473543446, + "cpu_time_us": 20.171, + "cuda_time_us": 8.131, + "pct_cuda_time": 0.14098366056419473, "trace": "" }, "children": [ @@ -12343,8 +12343,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.452, - "pct_cuda_time": 0.147264473543446, + "cuda_time_us": 8.131, + "pct_cuda_time": 0.14098366056419473, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -12354,9 +12354,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.769, - "cuda_time_us": 33.989, - "pct_cuda_time": 0.592211570192639, + "cpu_time_us": 75.533, + "cuda_time_us": 35.482, + "pct_cuda_time": 0.6152234957740447, "trace": "" }, "children": [ @@ -12364,8 +12364,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 33.989, - "pct_cuda_time": 0.592211570192639, + "cuda_time_us": 35.482, + "pct_cuda_time": 0.6152234957740447, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -12379,18 +12379,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 502.76, - "cuda_time_us": 163.022, - "pct_cuda_time": 2.8404340991480894, + "cpu_time_us": 571.565, + "cuda_time_us": 165.735, + "pct_cuda_time": 2.873684292658568, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.285, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cpu_time_us": 28.793, + "cuda_time_us": 4.887, + "pct_cuda_time": 0.08473584419840359, "trace": "" }, "children": [ @@ -12398,8 +12398,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cuda_time_us": 4.887, + "pct_cuda_time": 0.08473584419840359, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12409,18 +12409,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 295.06, - "cuda_time_us": 46.123000000000005, - "pct_cuda_time": 0.8036298288268292, + "cpu_time_us": 347.269, + "cuda_time_us": 49.516000000000005, + "pct_cuda_time": 0.8585594559705654, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 52.628, - "cuda_time_us": 16.634, - "pct_cuda_time": 0.2898245684952296, + "cpu_time_us": 51.176, + "cuda_time_us": 17.134, + "pct_cuda_time": 0.2970869561071101, "trace": "" }, "children": [ @@ -12428,8 +12428,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.634, - "pct_cuda_time": 0.2898245684952296, + "cuda_time_us": 17.134, + "pct_cuda_time": 0.2970869561071101, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12439,9 +12439,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.612, - "cuda_time_us": 3.684, - "pct_cuda_time": 0.0641886323395711, + "cpu_time_us": 28.102, + "cuda_time_us": 3.644, + "pct_cuda_time": 0.06318342874135108, "trace": "" }, "children": [ @@ -12449,8 +12449,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.684, - "pct_cuda_time": 0.0641886323395711, + "cuda_time_us": 3.644, + "pct_cuda_time": 0.06318342874135108, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -12460,9 +12460,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 135.111, - "cuda_time_us": 12.736, - "pct_cuda_time": 0.22190728053115574, + "cpu_time_us": 174.541, + "cuda_time_us": 15.139, + "pct_cuda_time": 0.26249558938400486, "trace": "" }, "children": [ @@ -12470,8 +12470,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.643, - "pct_cuda_time": 0.04605063932505061, + "cuda_time_us": 2.603, + "pct_cuda_time": 0.04513349753395633, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12480,8 +12480,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.128, - "pct_cuda_time": 0.10677197040632241, + "cuda_time_us": 8.732, + "pct_cuda_time": 0.15140441815847352, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12490,8 +12490,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12501,9 +12501,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 57.947, - "cuda_time_us": 13.069, - "pct_cuda_time": 0.22770934746087265, + "cpu_time_us": 58.468, + "cuda_time_us": 13.599, + "pct_cuda_time": 0.23579348173809914, "trace": "" }, "children": [ @@ -12511,8 +12511,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 13.069, - "pct_cuda_time": 0.22770934746087265, + "cuda_time_us": 13.599, + "pct_cuda_time": 0.23579348173809914, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12524,9 +12524,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 37.256, + "cpu_time_us": 31.437, "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -12535,7 +12535,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12545,18 +12545,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 129.774, - "cuda_time_us": 107.608, - "pct_cuda_time": 1.874921375894834, + "cpu_time_us": 143.204, + "cuda_time_us": 106.68599999999999, + "pct_cuda_time": 1.8498318547474697, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 50.255, - "cuda_time_us": 63.774, - "pct_cuda_time": 1.1111742233506536, + "cpu_time_us": 53.42, + "cuda_time_us": 63.845, + "pct_cuda_time": 1.1070104302940613, "trace": "" }, "children": [ @@ -12564,8 +12564,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.774, - "pct_cuda_time": 1.1111742233506536, + "cuda_time_us": 63.845, + "pct_cuda_time": 1.1070104302940613, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12575,9 +12575,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.647, - "cuda_time_us": 9.533, - "pct_cuda_time": 0.16609941153450905, + "cpu_time_us": 20.391, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "" }, "children": [ @@ -12585,8 +12585,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 9.533, - "pct_cuda_time": 0.16609941153450905, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -12596,9 +12596,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 47.28, - "cuda_time_us": 34.301, - "pct_cuda_time": 0.5976477410096712, + "cpu_time_us": 52.739, + "cuda_time_us": 34.55, + "pct_cuda_time": 0.59906351893899, "trace": "" }, "children": [ @@ -12606,8 +12606,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 34.301, - "pct_cuda_time": 0.5976477410096712, + "cuda_time_us": 34.55, + "pct_cuda_time": 0.59906351893899, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -12621,18 +12621,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 523.812, - "cuda_time_us": 163.43200000000002, - "pct_cuda_time": 2.847577785157651, + "cpu_time_us": 595.11, + "cuda_time_us": 162.77300000000002, + "pct_cuda_time": 2.8223260830175465, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.336, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.811, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "" }, "children": [ @@ -12640,8 +12640,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12651,18 +12651,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 321.43, - "cuda_time_us": 46.582, - "pct_cuda_time": 0.8116272724326551, + "cpu_time_us": 366.588, + "cuda_time_us": 48.126, + "pct_cuda_time": 0.8344582029654944, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.409, - "cuda_time_us": 17.264, - "pct_cuda_time": 0.3008014518757751, + "cpu_time_us": 54.231, + "cuda_time_us": 16.454, + "pct_cuda_time": 0.28529641506865827, "trace": "" }, "children": [ @@ -12670,8 +12670,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.264, - "pct_cuda_time": 0.3008014518757751, + "cuda_time_us": 16.454, + "pct_cuda_time": 0.28529641506865827, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12681,9 +12681,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.003, - "cuda_time_us": 3.844, - "pct_cuda_time": 0.0669764122457414, + "cpu_time_us": 23.465, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "" }, "children": [ @@ -12691,8 +12691,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.844, - "pct_cuda_time": 0.0669764122457414, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -12702,9 +12702,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 167.27, - "cuda_time_us": 12.776, - "pct_cuda_time": 0.22260422550769826, + "cpu_time_us": 195.362, + "cuda_time_us": 14.298, + "pct_cuda_time": 0.24791346436439013, "trace": "" }, "children": [ @@ -12712,8 +12712,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.563, - "pct_cuda_time": 0.04465674937196546, + "cuda_time_us": 2.683, + "pct_cuda_time": 0.04652062000906831, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12722,8 +12722,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.088, - "pct_cuda_time": 0.10607502542977984, + "cuda_time_us": 7.73, + "pct_cuda_time": 0.13403070915769588, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12732,8 +12732,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.125, - "pct_cuda_time": 0.071872450705953, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12743,9 +12743,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 58.697, - "cuda_time_us": 12.698, - "pct_cuda_time": 0.22124518280344027, + "cpu_time_us": 61.131, + "cuda_time_us": 13.69, + "pct_cuda_time": 0.237371333553539, "trace": "" }, "children": [ @@ -12753,8 +12753,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 12.698, - "pct_cuda_time": 0.22124518280344027, + "cuda_time_us": 13.69, + "pct_cuda_time": 0.237371333553539, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12766,9 +12766,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.209, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cpu_time_us": 31.487, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -12776,8 +12776,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12787,18 +12787,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 135.462, - "cuda_time_us": 107.47800000000001, - "pct_cuda_time": 1.8726563047210705, + "cpu_time_us": 148.512, + "cuda_time_us": 105.35500000000002, + "pct_cuda_time": 1.8267536045677946, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.851, - "cuda_time_us": 63.774, - "pct_cuda_time": 1.1111742233506536, + "cpu_time_us": 54.381, + "cuda_time_us": 63.404, + "pct_cuda_time": 1.0993639176500065, "trace": "" }, "children": [ @@ -12806,8 +12806,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.774, - "pct_cuda_time": 1.1111742233506536, + "cuda_time_us": 63.404, + "pct_cuda_time": 1.0993639176500065, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12817,9 +12817,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.777, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cpu_time_us": 22.634, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "" }, "children": [ @@ -12827,8 +12827,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -12838,9 +12838,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 50.906, - "cuda_time_us": 35.573, - "pct_cuda_time": 0.6198105912637251, + "cpu_time_us": 54.021, + "cuda_time_us": 33.74, + "pct_cuda_time": 0.5850189038784812, "trace": "" }, "children": [ @@ -12848,8 +12848,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 35.573, - "pct_cuda_time": 0.6198105912637251, + "cuda_time_us": 33.74, + "pct_cuda_time": 0.5850189038784812, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -12863,18 +12863,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 526.496, - "cuda_time_us": 159.689, - "pct_cuda_time": 2.7823611589776793, + "cpu_time_us": 628.329, + "cuda_time_us": 165.544, + "pct_cuda_time": 2.870372537749238, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.436, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cpu_time_us": 33.751, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -12882,8 +12882,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -12893,18 +12893,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 331.625, - "cuda_time_us": 48.17699999999999, - "pct_cuda_time": 0.8394179533722901, + "cpu_time_us": 377.855, + "cuda_time_us": 48.394999999999996, + "pct_cuda_time": 0.8391224022880583, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.539, - "cuda_time_us": 17.445, - "pct_cuda_time": 0.3039551278946303, + "cpu_time_us": 54.472, + "cuda_time_us": 17.605, + "pct_cuda_time": 0.305253639679332, "trace": "" }, "children": [ @@ -12912,8 +12912,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.445, - "pct_cuda_time": 0.3039551278946303, + "cuda_time_us": 17.605, + "pct_cuda_time": 0.305253639679332, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -12923,9 +12923,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.352, + "cpu_time_us": 26.41, "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "pct_cuda_time": 0.06666857396006995, "trace": "" }, "children": [ @@ -12934,7 +12934,7 @@ "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "pct_cuda_time": 0.06666857396006995, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -12944,9 +12944,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 144.165, - "cuda_time_us": 12.616999999999999, - "pct_cuda_time": 0.21983386922594153, + "cpu_time_us": 178.416, + "cuda_time_us": 14.057, + "pct_cuda_time": 0.24373475790811527, "trace": "" }, "children": [ @@ -12954,8 +12954,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.844, - "pct_cuda_time": 0.049552787832177043, + "cuda_time_us": 2.923, + "pct_cuda_time": 0.050681987434404274, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12964,8 +12964,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.728, - "pct_cuda_time": 0.09980252064089666, + "cuda_time_us": 7.41, + "pct_cuda_time": 0.12848221925724793, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12974,8 +12974,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.724, + "pct_cuda_time": 0.06457055121646307, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -12985,9 +12985,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 57.076, - "cuda_time_us": 14.27, - "pct_cuda_time": 0.24863512038156343, + "cpu_time_us": 65.108, + "cuda_time_us": 12.888, + "pct_cuda_time": 0.22346543074054132, "trace": "" }, "children": [ @@ -12995,8 +12995,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.27, - "pct_cuda_time": 0.24863512038156343, + "cuda_time_us": 12.888, + "pct_cuda_time": 0.22346543074054132, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13008,9 +13008,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.159, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 31.286, + "cuda_time_us": 4.886, + "pct_cuda_time": 0.0847185051674647, "trace": "" }, "children": [ @@ -13018,8 +13018,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.886, + "pct_cuda_time": 0.0847185051674647, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13029,18 +13029,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 127.089, - "cuda_time_us": 102.02000000000001, - "pct_cuda_time": 1.7775581626718364, + "cpu_time_us": 164.446, + "cuda_time_us": 107.537, + "pct_cuda_time": 1.864587370076474, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.371, - "cuda_time_us": 61.871, - "pct_cuda_time": 1.0780170660916406, + "cpu_time_us": 68.773, + "cuda_time_us": 63.814, + "pct_cuda_time": 1.1064729203349555, "trace": "" }, "children": [ @@ -13048,8 +13048,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.871, - "pct_cuda_time": 1.0780170660916406, + "cuda_time_us": 63.814, + "pct_cuda_time": 1.1064729203349555, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13059,9 +13059,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.096, - "cuda_time_us": 8.132, - "pct_cuda_time": 0.14168891373110537, + "cpu_time_us": 24.066, + "cuda_time_us": 8.332, + "pct_cuda_time": 0.14446880578291363, "trace": "" }, "children": [ @@ -13069,8 +13069,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.132, - "pct_cuda_time": 0.14168891373110537, + "cuda_time_us": 8.332, + "pct_cuda_time": 0.14446880578291363, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -13080,9 +13080,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 47.21, - "cuda_time_us": 32.017, - "pct_cuda_time": 0.5578521828490902, + "cpu_time_us": 54.291, + "cuda_time_us": 35.391, + "pct_cuda_time": 0.6136456439586048, "trace": "" }, "children": [ @@ -13090,8 +13090,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 32.017, - "pct_cuda_time": 0.5578521828490902, + "cuda_time_us": 35.391, + "pct_cuda_time": 0.6136456439586048, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -13105,18 +13105,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 476.361, - "cuda_time_us": 165.796, - "pct_cuda_time": 2.888767233271317, + "cpu_time_us": 557.133, + "cuda_time_us": 165.613, + "pct_cuda_time": 2.8715689308840218, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.755, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cpu_time_us": 29.324, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -13124,8 +13124,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13135,18 +13135,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 284.374, - "cuda_time_us": 48.727000000000004, - "pct_cuda_time": 0.8490009467997507, + "cpu_time_us": 337.064, + "cuda_time_us": 50.167, + "pct_cuda_time": 0.8698471651117892, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.169, - "cuda_time_us": 17.234, - "pct_cuda_time": 0.3002787431433682, + "cpu_time_us": 54.541, + "cuda_time_us": 17.435, + "pct_cuda_time": 0.30230600441971894, "trace": "" }, "children": [ @@ -13154,8 +13154,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.234, - "pct_cuda_time": 0.3002787431433682, + "cuda_time_us": 17.435, + "pct_cuda_time": 0.30230600441971894, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13165,9 +13165,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.702, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cpu_time_us": 24.827, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "" }, "children": [ @@ -13175,8 +13175,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -13186,9 +13186,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 135.322, - "cuda_time_us": 13.217, - "pct_cuda_time": 0.23028804387408017, + "cpu_time_us": 167.29, + "cuda_time_us": 14.257, + "pct_cuda_time": 0.24720256409589522, "trace": "" }, "children": [ @@ -13196,8 +13196,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.643, - "pct_cuda_time": 0.04605063932505061, + "cuda_time_us": 2.642, + "pct_cuda_time": 0.045809719740573414, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13206,8 +13206,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.529, - "pct_cuda_time": 0.11375884379616172, + "cuda_time_us": 7.73, + "pct_cuda_time": 0.13403070915769588, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13216,8 +13216,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13227,9 +13227,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 54.211, - "cuda_time_us": 14.311, - "pct_cuda_time": 0.2493494889825196, + "cpu_time_us": 57.876, + "cuda_time_us": 14.63, + "pct_cuda_time": 0.2536700226361049, "trace": "" }, "children": [ @@ -13237,8 +13237,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.311, - "pct_cuda_time": 0.2493494889825196, + "cuda_time_us": 14.63, + "pct_cuda_time": 0.2536700226361049, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13250,9 +13250,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.248, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cpu_time_us": 28.112, + "cuda_time_us": 5.046, + "pct_cuda_time": 0.08749275011768867, "trace": "" }, "children": [ @@ -13260,8 +13260,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cuda_time_us": 5.046, + "pct_cuda_time": 0.08749275011768867, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13271,18 +13271,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 122.643, - "cuda_time_us": 107.618, - "pct_cuda_time": 1.8750956121389692, + "cpu_time_us": 141.292, + "cuda_time_us": 105.714, + "pct_cuda_time": 1.8329783166748592, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 45.659, - "cuda_time_us": 62.272, - "pct_cuda_time": 1.0850039394814799, + "cpu_time_us": 50.055, + "cuda_time_us": 61.921, + "pct_cuda_time": 1.073650134767618, "trace": "" }, "children": [ @@ -13290,8 +13290,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.272, - "pct_cuda_time": 1.0850039394814799, + "cuda_time_us": 61.921, + "pct_cuda_time": 1.073650134767618, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13301,9 +13301,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.056, - "cuda_time_us": 8.291, - "pct_cuda_time": 0.14445927001286213, + "cpu_time_us": 21.802, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "" }, "children": [ @@ -13311,8 +13311,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.291, - "pct_cuda_time": 0.14445927001286213, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -13322,9 +13322,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.378, - "cuda_time_us": 37.055, - "pct_cuda_time": 0.6456324026446274, + "cpu_time_us": 53.039, + "cuda_time_us": 35.582, + "pct_cuda_time": 0.6169573988679347, "trace": "" }, "children": [ @@ -13332,8 +13332,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 37.055, - "pct_cuda_time": 0.6456324026446274, + "cuda_time_us": 35.582, + "pct_cuda_time": 0.6169573988679347, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -13347,18 +13347,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 467.277, - "cuda_time_us": 166.17900000000003, - "pct_cuda_time": 2.895440481421713, + "cpu_time_us": 564.263, + "cuda_time_us": 166.36900000000003, + "pct_cuda_time": 2.884677238273831, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.045, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 26.479, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -13366,8 +13366,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13377,18 +13377,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 280.098, - "cuda_time_us": 49.229, - "pct_cuda_time": 0.85774760625536, + "cpu_time_us": 343.504, + "cuda_time_us": 50.781000000000006, + "pct_cuda_time": 0.8804933301082736, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.509, - "cuda_time_us": 17.635, - "pct_cuda_time": 0.30726561653320755, + "cpu_time_us": 51.627, + "cuda_time_us": 17.986, + "pct_cuda_time": 0.3118598104670528, "trace": "" }, "children": [ @@ -13396,8 +13396,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.635, - "pct_cuda_time": 0.30726561653320755, + "cuda_time_us": 17.986, + "pct_cuda_time": 0.3118598104670528, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13407,9 +13407,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.432, + "cpu_time_us": 33.36, "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "pct_cuda_time": 0.06597501272251395, "trace": "" }, "children": [ @@ -13418,7 +13418,7 @@ "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "pct_cuda_time": 0.06597501272251395, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -13428,9 +13428,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 132.468, - "cuda_time_us": 12.897000000000002, - "pct_cuda_time": 0.2247124840617396, + "cpu_time_us": 170.275, + "cuda_time_us": 15.26, + "pct_cuda_time": 0.2645936121276118, "trace": "" }, "children": [ @@ -13438,8 +13438,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.723, - "pct_cuda_time": 0.04744452927813575, + "cuda_time_us": 2.683, + "pct_cuda_time": 0.04652062000906831, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13448,8 +13448,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.089, - "pct_cuda_time": 0.1060924490541934, + "cuda_time_us": 8.812, + "pct_cuda_time": 0.15279154063358552, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13458,8 +13458,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.085, - "pct_cuda_time": 0.07117550572941042, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13469,9 +13469,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 52.839, - "cuda_time_us": 14.892, - "pct_cuda_time": 0.25947261476680045, + "cpu_time_us": 57.777, + "cuda_time_us": 13.73, + "pct_cuda_time": 0.23806489479109502, "trace": "" }, "children": [ @@ -13479,8 +13479,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.892, - "pct_cuda_time": 0.25947261476680045, + "cuda_time_us": 13.73, + "pct_cuda_time": 0.23806489479109502, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13492,9 +13492,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.817, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 29.104, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "" }, "children": [ @@ -13502,8 +13502,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13513,18 +13513,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 122.874, - "cuda_time_us": 107.81800000000001, - "pct_cuda_time": 1.8785803370216827, + "cpu_time_us": 141.581, + "cuda_time_us": 106.21600000000001, + "pct_cuda_time": 1.8416825102061871, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 46.289, - "cuda_time_us": 62.322, - "pct_cuda_time": 1.0858751207021582, + "cpu_time_us": 51.437, + "cuda_time_us": 62.693, + "pct_cuda_time": 1.0870358666524487, "trace": "" }, "children": [ @@ -13532,8 +13532,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.322, - "pct_cuda_time": 1.0858751207021582, + "cuda_time_us": 62.693, + "pct_cuda_time": 1.0870358666524487, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13543,9 +13543,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.025, - "cuda_time_us": 9.253, - "pct_cuda_time": 0.16122079669871103, + "cpu_time_us": 18.958, + "cuda_time_us": 8.251, + "pct_cuda_time": 0.1430643442768627, "trace": "" }, "children": [ @@ -13553,8 +13553,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 9.253, - "pct_cuda_time": 0.16122079669871103, + "cuda_time_us": 8.251, + "pct_cuda_time": 0.1430643442768627, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -13564,9 +13564,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.628, - "cuda_time_us": 36.243, - "pct_cuda_time": 0.6314844196208133, + "cpu_time_us": 51.878, + "cuda_time_us": 35.272, + "pct_cuda_time": 0.6115822992768757, "trace": "" }, "children": [ @@ -13574,8 +13574,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 36.243, - "pct_cuda_time": 0.6314844196208133, + "cuda_time_us": 35.272, + "pct_cuda_time": 0.6115822992768757, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -13589,18 +13589,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 554.357, - "cuda_time_us": 161.50900000000001, - "pct_cuda_time": 2.8140721554103667, + "cpu_time_us": 636.872, + "cuda_time_us": 164.114, + "pct_cuda_time": 2.845577723506611, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 22.614, - "cuda_time_us": 4.405, - "pct_cuda_time": 0.07675106554175103, + "cpu_time_us": 28.082, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "" }, "children": [ @@ -13608,8 +13608,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.405, - "pct_cuda_time": 0.07675106554175103, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13619,18 +13619,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 363.232, - "cuda_time_us": 47.735, - "pct_cuda_time": 0.8317167113814947, + "cpu_time_us": 413.888, + "cuda_time_us": 50.448, + "pct_cuda_time": 0.87471943280562, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 44.887, - "cuda_time_us": 16.724, - "pct_cuda_time": 0.2913926946924504, + "cpu_time_us": 51.537, + "cuda_time_us": 17.345, + "pct_cuda_time": 0.30074549163521797, "trace": "" }, "children": [ @@ -13638,8 +13638,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.724, - "pct_cuda_time": 0.2913926946924504, + "cuda_time_us": 17.345, + "pct_cuda_time": 0.30074549163521797, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13649,9 +13649,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.381, - "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "cpu_time_us": 25.387, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "" }, "children": [ @@ -13659,8 +13659,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -13670,9 +13670,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 208.181, - "cuda_time_us": 12.495, - "pct_cuda_time": 0.2177081870474867, + "cpu_time_us": 237.916, + "cuda_time_us": 14.779, + "pct_cuda_time": 0.256253538246001, "trace": "" }, "children": [ @@ -13680,8 +13680,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.563, - "pct_cuda_time": 0.04465674937196546, + "cuda_time_us": 2.522, + "pct_cuda_time": 0.043729036027905434, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13690,8 +13690,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.887, - "pct_cuda_time": 0.10257287692265339, + "cuda_time_us": 8.332, + "pct_cuda_time": 0.14446880578291363, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13700,8 +13700,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13711,9 +13711,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 60.3, - "cuda_time_us": 14.711, - "pct_cuda_time": 0.25631893874794537, + "cpu_time_us": 65.198, + "cuda_time_us": 14.64, + "pct_cuda_time": 0.2538434129454939, "trace": "" }, "children": [ @@ -13721,8 +13721,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.711, - "pct_cuda_time": 0.25631893874794537, + "cuda_time_us": 14.64, + "pct_cuda_time": 0.2538434129454939, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13734,9 +13734,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 27.071, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cpu_time_us": 28.443, + "cuda_time_us": 5.287, + "pct_cuda_time": 0.09167145657396354, "trace": "" }, "children": [ @@ -13744,8 +13744,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cuda_time_us": 5.287, + "pct_cuda_time": 0.09167145657396354, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13755,18 +13755,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 125.097, - "cuda_time_us": 104.72300000000001, - "pct_cuda_time": 1.8246542194617008, + "cpu_time_us": 147.57, + "cuda_time_us": 103.773, + "pct_cuda_time": 1.7993232576224545, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 46.259, - "cuda_time_us": 62.382, - "pct_cuda_time": 1.086920538166972, + "cpu_time_us": 52.789, + "cuda_time_us": 62.833, + "pct_cuda_time": 1.0894633309838946, "trace": "" }, "children": [ @@ -13774,8 +13774,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.382, - "pct_cuda_time": 1.086920538166972, + "cuda_time_us": 62.833, + "pct_cuda_time": 1.0894633309838946, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13785,9 +13785,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.867, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.1674933014875942, + "cpu_time_us": 24.006, + "cuda_time_us": 8.212, + "pct_cuda_time": 0.14238812207024562, "trace": "" }, "children": [ @@ -13795,8 +13795,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 9.613, - "pct_cuda_time": 0.1674933014875942, + "cuda_time_us": 8.212, + "pct_cuda_time": 0.14238812207024562, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -13806,9 +13806,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.449, + "cpu_time_us": 53.55, "cuda_time_us": 32.728, - "pct_cuda_time": 0.5702403798071344, + "pct_cuda_time": 0.5674718045683145, "trace": "" }, "children": [ @@ -13817,7 +13817,7 @@ "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, "cuda_time_us": 32.728, - "pct_cuda_time": 0.5702403798071344, + "pct_cuda_time": 0.5674718045683145, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -13831,18 +13831,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 480.728, - "cuda_time_us": 159.458, - "pct_cuda_time": 2.778336301738146, + "cpu_time_us": 556.853, + "cuda_time_us": 163.36400000000003, + "pct_cuda_time": 2.8325734503024367, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.555, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "cpu_time_us": 28.032, + "cuda_time_us": 4.886, + "pct_cuda_time": 0.0847185051674647, "trace": "" }, "children": [ @@ -13850,8 +13850,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "cuda_time_us": 4.886, + "pct_cuda_time": 0.0847185051674647, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13861,18 +13861,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 288.711, - "cuda_time_us": 48.397, - "pct_cuda_time": 0.8432511507432744, + "cpu_time_us": 334.15, + "cuda_time_us": 48.707, + "pct_cuda_time": 0.8445321799409954, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.179, - "cuda_time_us": 17.726, - "pct_cuda_time": 0.30885116635484183, + "cpu_time_us": 54.932, + "cuda_time_us": 17.515, + "pct_cuda_time": 0.30369312689483097, "trace": "" }, "children": [ @@ -13880,8 +13880,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.726, - "pct_cuda_time": 0.30885116635484183, + "cuda_time_us": 17.515, + "pct_cuda_time": 0.30369312689483097, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13891,9 +13891,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.654, - "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "cpu_time_us": 24.707, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "" }, "children": [ @@ -13901,8 +13901,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -13912,9 +13912,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 135.082, - "cuda_time_us": 12.736, - "pct_cuda_time": 0.22190728053115574, + "cpu_time_us": 164.646, + "cuda_time_us": 13.978, + "pct_cuda_time": 0.24236497446394217, "trace": "" }, "children": [ @@ -13923,7 +13923,7 @@ "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, "cuda_time_us": 2.683, - "pct_cuda_time": 0.04674758430159318, + "pct_cuda_time": 0.04652062000906831, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13932,8 +13932,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.088, - "pct_cuda_time": 0.10607502542977984, + "cuda_time_us": 7.61, + "pct_cuda_time": 0.1319500254450279, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13942,8 +13942,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.685, + "pct_cuda_time": 0.06389432900984597, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -13953,9 +13953,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 56.945, - "cuda_time_us": 14.13, - "pct_cuda_time": 0.24619581296366444, + "cpu_time_us": 57.626, + "cuda_time_us": 13.449, + "pct_cuda_time": 0.23319262709726418, "trace": "" }, "children": [ @@ -13963,8 +13963,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.13, - "pct_cuda_time": 0.24619581296366444, + "cuda_time_us": 13.449, + "pct_cuda_time": 0.23319262709726418, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -13976,9 +13976,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.029, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.08025321404887745, + "cpu_time_us": 26.75, + "cuda_time_us": 4.727, + "pct_cuda_time": 0.08196159924817963, "trace": "" }, "children": [ @@ -13986,8 +13986,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.08025321404887745, + "cuda_time_us": 4.727, + "pct_cuda_time": 0.08196159924817963, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -13997,18 +13997,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 125.868, - "cuda_time_us": 101.729, - "pct_cuda_time": 1.7724878879674888, + "cpu_time_us": 146.96, + "cuda_time_us": 105.04400000000001, + "pct_cuda_time": 1.8213611659457964, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 46.329, - "cuda_time_us": 62.142, - "pct_cuda_time": 1.0827388683077166, + "cpu_time_us": 53.21, + "cuda_time_us": 63.364, + "pct_cuda_time": 1.0986703564124505, "trace": "" }, "children": [ @@ -14016,8 +14016,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.142, - "pct_cuda_time": 1.0827388683077166, + "cuda_time_us": 63.364, + "pct_cuda_time": 1.0986703564124505, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14027,9 +14027,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 18.528, - "cuda_time_us": 8.171, - "pct_cuda_time": 0.14236843508323438, + "cpu_time_us": 21.612, + "cuda_time_us": 8.131, + "pct_cuda_time": 0.14098366056419473, "trace": "" }, "children": [ @@ -14037,8 +14037,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.171, - "pct_cuda_time": 0.14236843508323438, + "cuda_time_us": 8.131, + "pct_cuda_time": 0.14098366056419473, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -14048,9 +14048,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.199, - "cuda_time_us": 31.416, - "pct_cuda_time": 0.5473805845765379, + "cpu_time_us": 51.197, + "cuda_time_us": 33.549, + "pct_cuda_time": 0.5817071489691512, "trace": "" }, "children": [ @@ -14058,8 +14058,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 31.416, - "pct_cuda_time": 0.5473805845765379, + "cuda_time_us": 33.549, + "pct_cuda_time": 0.5817071489691512, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -14073,18 +14073,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 471.955, - "cuda_time_us": 168.769, - "pct_cuda_time": 2.940567668652844, + "cpu_time_us": 571.123, + "cuda_time_us": 165.005, + "pct_cuda_time": 2.8610268000731707, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 22.935, + "cpu_time_us": 27.281, "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -14093,7 +14093,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14103,18 +14103,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 284.725, - "cuda_time_us": 48.865, - "pct_cuda_time": 0.8514054069688226, + "cpu_time_us": 338.446, + "cuda_time_us": 48.95700000000001, + "pct_cuda_time": 0.8488669376757204, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 45.197, - "cuda_time_us": 17.725, - "pct_cuda_time": 0.30883374273042835, + "cpu_time_us": 51.127, + "cuda_time_us": 17.315, + "pct_cuda_time": 0.30022532070705105, "trace": "" }, "children": [ @@ -14122,8 +14122,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.725, - "pct_cuda_time": 0.30883374273042835, + "cuda_time_us": 17.315, + "pct_cuda_time": 0.30022532070705105, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14133,9 +14133,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 24.697, - "cuda_time_us": 3.884, - "pct_cuda_time": 0.06767335722228397, + "cpu_time_us": 24.527, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "" }, "children": [ @@ -14143,8 +14143,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.884, - "pct_cuda_time": 0.06767335722228397, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -14154,9 +14154,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 134.511, - "cuda_time_us": 13.136, - "pct_cuda_time": 0.22887673029658145, + "cpu_time_us": 171.176, + "cuda_time_us": 14.779, + "pct_cuda_time": 0.256253538246001, "trace": "" }, "children": [ @@ -14164,8 +14164,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.803, - "pct_cuda_time": 0.048838419231220905, + "cuda_time_us": 2.883, + "pct_cuda_time": 0.04998842619684828, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14174,8 +14174,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.248, - "pct_cuda_time": 0.10886280533595014, + "cuda_time_us": 7.971, + "pct_cuda_time": 0.13820941561397074, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14184,8 +14184,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.085, - "pct_cuda_time": 0.07117550572941042, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14195,9 +14195,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 53.99, - "cuda_time_us": 14.12, - "pct_cuda_time": 0.24602157671952876, + "cpu_time_us": 58.888, + "cuda_time_us": 12.938, + "pct_cuda_time": 0.22433238228748634, "trace": "" }, "children": [ @@ -14205,8 +14205,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.12, - "pct_cuda_time": 0.24602157671952876, + "cuda_time_us": 12.938, + "pct_cuda_time": 0.22433238228748634, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14218,9 +14218,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.406, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "cpu_time_us": 28.954, + "cuda_time_us": 5.367, + "pct_cuda_time": 0.09305857904907552, "trace": "" }, "children": [ @@ -14228,8 +14228,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "cuda_time_us": 5.367, + "pct_cuda_time": 0.09305857904907552, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14239,18 +14239,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 124.005, - "cuda_time_us": 110.452, - "pct_cuda_time": 1.9244741637270106, + "cpu_time_us": 155.373, + "cuda_time_us": 105.955, + "pct_cuda_time": 1.8371570231311343, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 46.209, - "cuda_time_us": 64.545, - "pct_cuda_time": 1.1246078377735118, + "cpu_time_us": 57.196, + "cuda_time_us": 62.031, + "pct_cuda_time": 1.0755574281708968, "trace": "" }, "children": [ @@ -14258,8 +14258,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 64.545, - "pct_cuda_time": 1.1246078377735118, + "cuda_time_us": 62.031, + "pct_cuda_time": 1.0755574281708968, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14269,9 +14269,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.586, - "cuda_time_us": 8.452, - "pct_cuda_time": 0.147264473543446, + "cpu_time_us": 22.514, + "cuda_time_us": 8.252, + "pct_cuda_time": 0.14308168330780163, "trace": "" }, "children": [ @@ -14279,8 +14279,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.452, - "pct_cuda_time": 0.147264473543446, + "cuda_time_us": 8.252, + "pct_cuda_time": 0.14308168330780163, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -14290,9 +14290,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.239, - "cuda_time_us": 37.455, - "pct_cuda_time": 0.6526018524100532, + "cpu_time_us": 55.392, + "cuda_time_us": 35.672, + "pct_cuda_time": 0.6185179116524356, "trace": "" }, "children": [ @@ -14300,8 +14300,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 37.455, - "pct_cuda_time": 0.6526018524100532, + "cuda_time_us": 35.672, + "pct_cuda_time": 0.6185179116524356, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -14315,18 +14315,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 492.114, - "cuda_time_us": 162.19099999999997, - "pct_cuda_time": 2.825955067260417, + "cpu_time_us": 583.933, + "cuda_time_us": 161.76100000000002, + "pct_cuda_time": 2.8047789837073798, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.717, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.08025321404887745, + "cpu_time_us": 31.557, + "cuda_time_us": 5.207, + "pct_cuda_time": 0.09028433409885155, "trace": "" }, "children": [ @@ -14334,8 +14334,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.08025321404887745, + "cuda_time_us": 5.207, + "pct_cuda_time": 0.09028433409885155, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14345,18 +14345,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 281.1, - "cuda_time_us": 48.756, - "pct_cuda_time": 0.8495062319077441, + "cpu_time_us": 356.623, + "cuda_time_us": 46.373000000000005, + "pct_cuda_time": 0.804062881729603, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 45.678, - "cuda_time_us": 17.274, - "pct_cuda_time": 0.3009756881199108, + "cpu_time_us": 53.53, + "cuda_time_us": 17.064, + "pct_cuda_time": 0.29587322394138715, "trace": "" }, "children": [ @@ -14364,8 +14364,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.274, - "pct_cuda_time": 0.3009756881199108, + "cuda_time_us": 17.064, + "pct_cuda_time": 0.29587322394138715, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14375,9 +14375,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.833, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cpu_time_us": 28.543, + "cuda_time_us": 3.725, + "pct_cuda_time": 0.06458789024740197, "trace": "" }, "children": [ @@ -14385,8 +14385,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cuda_time_us": 3.725, + "pct_cuda_time": 0.06458789024740197, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -14396,9 +14396,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 133.759, - "cuda_time_us": 12.936, - "pct_cuda_time": 0.22539200541386856, + "cpu_time_us": 179.518, + "cuda_time_us": 11.855, + "pct_cuda_time": 0.2055542117806578, "trace": "" }, "children": [ @@ -14406,8 +14406,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.763, - "pct_cuda_time": 0.04814147425467833, + "cuda_time_us": 2.883, + "pct_cuda_time": 0.04998842619684828, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14416,8 +14416,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.208, - "pct_cuda_time": 0.10816586035940756, + "cuda_time_us": 6.008, + "pct_cuda_time": 0.10417289788091032, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14426,8 +14426,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 2.964, + "pct_cuda_time": 0.051392887702899175, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14437,9 +14437,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 53.881, - "cuda_time_us": 14.701, - "pct_cuda_time": 0.25614470250380966, + "cpu_time_us": 59.499, + "cuda_time_us": 13.729, + "pct_cuda_time": 0.23804755576015613, "trace": "" }, "children": [ @@ -14447,8 +14447,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.701, - "pct_cuda_time": 0.25614470250380966, + "cuda_time_us": 13.729, + "pct_cuda_time": 0.23804755576015613, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14460,9 +14460,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 35.113, - "cuda_time_us": 5.047, - "pct_cuda_time": 0.08793703241525933, + "cpu_time_us": 29.664, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -14470,8 +14470,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.047, - "pct_cuda_time": 0.08793703241525933, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14481,18 +14481,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 135.362, - "cuda_time_us": 103.78199999999998, - "pct_cuda_time": 1.8082585888885363, + "cpu_time_us": 147.23, + "cuda_time_us": 105.495, + "pct_cuda_time": 1.8291810688992403, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 54.892, - "cuda_time_us": 62.022, - "pct_cuda_time": 1.0806480333780888, + "cpu_time_us": 52.368, + "cuda_time_us": 63.043, + "pct_cuda_time": 1.0931045274810636, "trace": "" }, "children": [ @@ -14500,8 +14500,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.022, - "pct_cuda_time": 1.0806480333780888, + "cuda_time_us": 63.043, + "pct_cuda_time": 1.0931045274810636, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14511,9 +14511,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.997, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cpu_time_us": 22.463, + "cuda_time_us": 8.172, + "pct_cuda_time": 0.14169456083268964, "trace": "" }, "children": [ @@ -14521,8 +14521,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cuda_time_us": 8.172, + "pct_cuda_time": 0.14169456083268964, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -14532,9 +14532,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.179, - "cuda_time_us": 33.629, - "pct_cuda_time": 0.585939065403756, + "cpu_time_us": 52.729, + "cuda_time_us": 34.28, + "pct_cuda_time": 0.5943819805854871, "trace": "" }, "children": [ @@ -14542,8 +14542,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 33.629, - "pct_cuda_time": 0.585939065403756, + "cuda_time_us": 34.28, + "pct_cuda_time": 0.5943819805854871, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -14557,18 +14557,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 469.631, - "cuda_time_us": 164.523, - "pct_cuda_time": 2.8665869593928495, + "cpu_time_us": 581.94, + "cuda_time_us": 163.132, + "pct_cuda_time": 2.8285507951246114, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.174, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.0788593240957923, + "cpu_time_us": 28.833, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -14576,8 +14576,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.526, - "pct_cuda_time": 0.0788593240957923, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14587,18 +14587,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 280.92, - "cuda_time_us": 48.19500000000001, - "pct_cuda_time": 0.8397315786117345, + "cpu_time_us": 354.33, + "cuda_time_us": 48.035, + "pct_cuda_time": 0.8328803511500544, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 44.837, - "cuda_time_us": 17.344, - "pct_cuda_time": 0.3021953418288603, + "cpu_time_us": 52.919, + "cuda_time_us": 17.175, + "pct_cuda_time": 0.297797856375605, "trace": "" }, "children": [ @@ -14606,8 +14606,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.344, - "pct_cuda_time": 0.3021953418288603, + "cuda_time_us": 17.175, + "pct_cuda_time": 0.297797856375605, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14617,9 +14617,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.762, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cpu_time_us": 25.368, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "" }, "children": [ @@ -14627,8 +14627,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -14638,9 +14638,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 133.509, - "cuda_time_us": 12.696000000000002, - "pct_cuda_time": 0.22121033555461317, + "cpu_time_us": 172.248, + "cuda_time_us": 14.058000000000002, + "pct_cuda_time": 0.2437520969390542, "trace": "" }, "children": [ @@ -14649,7 +14649,7 @@ "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, "cuda_time_us": 2.723, - "pct_cuda_time": 0.04744452927813575, + "pct_cuda_time": 0.047214181246624305, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14658,8 +14658,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.048, - "pct_cuda_time": 0.10537808045323727, + "cuda_time_us": 7.49, + "pct_cuda_time": 0.12986934173235992, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14668,8 +14668,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.925, - "pct_cuda_time": 0.06838772582324011, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14679,9 +14679,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 53.36, - "cuda_time_us": 14.31, - "pct_cuda_time": 0.249332065358106, + "cpu_time_us": 72.138, + "cuda_time_us": 12.998, + "pct_cuda_time": 0.2253727241438203, "trace": "" }, "children": [ @@ -14689,8 +14689,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.31, - "pct_cuda_time": 0.249332065358106, + "cuda_time_us": 12.998, + "pct_cuda_time": 0.2253727241438203, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14702,9 +14702,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.747, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cpu_time_us": 29.224, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -14712,8 +14712,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14723,18 +14723,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 122.614, - "cuda_time_us": 106.916, - "pct_cuda_time": 1.8628642278006473, + "cpu_time_us": 147.521, + "cuda_time_us": 105.685, + "pct_cuda_time": 1.8324754847776312, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.261, - "cuda_time_us": 61.801, - "pct_cuda_time": 1.076797412382691, + "cpu_time_us": 57.335, + "cuda_time_us": 63.804, + "pct_cuda_time": 1.1062995300255665, "trace": "" }, "children": [ @@ -14742,8 +14742,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.801, - "pct_cuda_time": 1.076797412382691, + "cuda_time_us": 63.804, + "pct_cuda_time": 1.1062995300255665, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14753,9 +14753,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 16.675, - "cuda_time_us": 8.372, - "pct_cuda_time": 0.14587058359036084, + "cpu_time_us": 23.114, + "cuda_time_us": 8.212, + "pct_cuda_time": 0.14238812207024562, "trace": "" }, "children": [ @@ -14763,8 +14763,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.372, - "pct_cuda_time": 0.14587058359036084, + "cuda_time_us": 8.212, + "pct_cuda_time": 0.14238812207024562, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -14774,9 +14774,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 44.476, - "cuda_time_us": 36.743, - "pct_cuda_time": 0.6401962318275954, + "cpu_time_us": 49.594, + "cuda_time_us": 33.669, + "pct_cuda_time": 0.5837878326818192, "trace": "" }, "children": [ @@ -14784,8 +14784,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 36.743, - "pct_cuda_time": 0.6401962318275954, + "cuda_time_us": 33.669, + "pct_cuda_time": 0.5837878326818192, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -14799,18 +14799,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 462.14, - "cuda_time_us": 162.62199999999999, - "pct_cuda_time": 2.8334646493826634, + "cpu_time_us": 575.21, + "cuda_time_us": 162.042, + "pct_cuda_time": 2.8096512514012106, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.064, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.11, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "" }, "children": [ @@ -14818,8 +14818,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14829,18 +14829,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 276.382, - "cuda_time_us": 48.626999999999995, - "pct_cuda_time": 0.8472585843583942, + "cpu_time_us": 348.281, + "cuda_time_us": 47.134, + "pct_cuda_time": 0.8172578842741057, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 45.498, - "cuda_time_us": 17.595, - "pct_cuda_time": 0.3065686715566649, + "cpu_time_us": 53.92, + "cuda_time_us": 17.645, + "pct_cuda_time": 0.30594720091688793, "trace": "" }, "children": [ @@ -14848,8 +14848,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.595, - "pct_cuda_time": 0.3065686715566649, + "cuda_time_us": 17.645, + "pct_cuda_time": 0.30594720091688793, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14859,9 +14859,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 20.851, - "cuda_time_us": 3.725, - "pct_cuda_time": 0.06490300094052724, + "cpu_time_us": 25.117, + "cuda_time_us": 3.645, + "pct_cuda_time": 0.06320076777228997, "trace": "" }, "children": [ @@ -14869,8 +14869,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.725, - "pct_cuda_time": 0.06490300094052724, + "cuda_time_us": 3.645, + "pct_cuda_time": 0.06320076777228997, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -14880,9 +14880,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 130.995, - "cuda_time_us": 13.056, - "pct_cuda_time": 0.2274828403434963, + "cpu_time_us": 178.186, + "cuda_time_us": 13.617, + "pct_cuda_time": 0.23610558429499937, "trace": "" }, "children": [ @@ -14891,7 +14891,7 @@ "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, "cuda_time_us": 2.643, - "pct_cuda_time": 0.04605063932505061, + "pct_cuda_time": 0.045827058771512316, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14900,8 +14900,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.368, - "pct_cuda_time": 0.11095364026557787, + "cuda_time_us": 7.209, + "pct_cuda_time": 0.12499707403852905, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14910,8 +14910,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -14921,9 +14921,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 52.779, - "cuda_time_us": 14.251, - "pct_cuda_time": 0.2483040715177057, + "cpu_time_us": 59.338, + "cuda_time_us": 12.227, + "pct_cuda_time": 0.21200433128992854, "trace": "" }, "children": [ @@ -14931,8 +14931,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.251, - "pct_cuda_time": 0.2483040715177057, + "cuda_time_us": 12.227, + "pct_cuda_time": 0.21200433128992854, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14944,9 +14944,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.357, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 27.381, + "cuda_time_us": 5.047, + "pct_cuda_time": 0.08751008914862757, "trace": "" }, "children": [ @@ -14954,8 +14954,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 5.047, + "pct_cuda_time": 0.08751008914862757, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -14965,18 +14965,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 122.192, - "cuda_time_us": 104.743, - "pct_cuda_time": 1.8250026919499718, + "cpu_time_us": 153.499, + "cuda_time_us": 105.255, + "pct_cuda_time": 1.8250197014739042, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 45.488, - "cuda_time_us": 61.971, - "pct_cuda_time": 1.079759428532997, + "cpu_time_us": 61.732, + "cuda_time_us": 62.012, + "pct_cuda_time": 1.0752279865830578, "trace": "" }, "children": [ @@ -14984,8 +14984,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.971, - "pct_cuda_time": 1.079759428532997, + "cuda_time_us": 62.012, + "pct_cuda_time": 1.0752279865830578, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -14995,9 +14995,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.226, - "cuda_time_us": 8.612, - "pct_cuda_time": 0.15005225344961629, + "cpu_time_us": 21.593, + "cuda_time_us": 8.171, + "pct_cuda_time": 0.14167722180175069, "trace": "" }, "children": [ @@ -15005,8 +15005,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.612, - "pct_cuda_time": 0.15005225344961629, + "cuda_time_us": 8.171, + "pct_cuda_time": 0.14167722180175069, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -15016,9 +15016,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.568, - "cuda_time_us": 34.16, - "pct_cuda_time": 0.5951910099673585, + "cpu_time_us": 53.25, + "cuda_time_us": 35.072, + "pct_cuda_time": 0.6081144930890958, "trace": "" }, "children": [ @@ -15026,8 +15026,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 34.16, - "pct_cuda_time": 0.5951910099673585, + "cuda_time_us": 35.072, + "pct_cuda_time": 0.6081144930890958, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -15041,18 +15041,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 475.079, - "cuda_time_us": 167.35999999999999, - "pct_cuda_time": 2.9160177818541313, + "cpu_time_us": 568.34, + "cuda_time_us": 164.706, + "pct_cuda_time": 2.8558424298224394, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.085, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.411, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -15060,8 +15060,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15071,18 +15071,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 283.984, - "cuda_time_us": 47.79599999999999, - "pct_cuda_time": 0.8327795524707221, + "cpu_time_us": 348.141, + "cuda_time_us": 49.19799999999999, + "pct_cuda_time": 0.853045644131995, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.85, - "cuda_time_us": 16.945, - "pct_cuda_time": 0.2952433156878481, + "cpu_time_us": 53.179, + "cuda_time_us": 17.566, + "pct_cuda_time": 0.30457741747271483, "trace": "" }, "children": [ @@ -15090,8 +15090,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.945, - "pct_cuda_time": 0.2952433156878481, + "cuda_time_us": 17.566, + "pct_cuda_time": 0.30457741747271483, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15101,9 +15101,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.312, - "cuda_time_us": 3.804, - "pct_cuda_time": 0.06627946726919884, + "cpu_time_us": 24.627, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "" }, "children": [ @@ -15111,8 +15111,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.804, - "pct_cuda_time": 0.06627946726919884, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -15122,9 +15122,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 132.928, - "cuda_time_us": 12.857, - "pct_cuda_time": 0.224015539085197, + "cpu_time_us": 178.817, + "cuda_time_us": 13.456999999999999, + "pct_cuda_time": 0.23333133934477535, "trace": "" }, "children": [ @@ -15132,8 +15132,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.563, - "pct_cuda_time": 0.04465674937196546, + "cuda_time_us": 2.683, + "pct_cuda_time": 0.04652062000906831, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15142,8 +15142,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.369, - "pct_cuda_time": 0.11097106388999142, + "cuda_time_us": 6.889, + "pct_cuda_time": 0.11944858413808111, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15152,8 +15152,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.925, - "pct_cuda_time": 0.06838772582324011, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15163,9 +15163,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 56.615, - "cuda_time_us": 14.19, - "pct_cuda_time": 0.2472412304284783, + "cpu_time_us": 58.388, + "cuda_time_us": 14.41, + "pct_cuda_time": 0.24985543582954692, "trace": "" }, "children": [ @@ -15173,8 +15173,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.19, - "pct_cuda_time": 0.2472412304284783, + "cuda_time_us": 14.41, + "pct_cuda_time": 0.24985543582954692, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15186,9 +15186,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.428, - "cuda_time_us": 5.046, - "pct_cuda_time": 0.08791960879084577, + "cpu_time_us": 28.572, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -15196,8 +15196,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.046, - "pct_cuda_time": 0.08791960879084577, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15207,18 +15207,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 126.569, - "cuda_time_us": 109.952, - "pct_cuda_time": 1.9157623515202287, + "cpu_time_us": 142.834, + "cuda_time_us": 106.016, + "pct_cuda_time": 1.8382147040184074, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.611, - "cuda_time_us": 64.185, - "pct_cuda_time": 1.1183353329846286, + "cpu_time_us": 55.813, + "cuda_time_us": 62.993, + "pct_cuda_time": 1.0922375759341187, "trace": "" }, "children": [ @@ -15226,8 +15226,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 64.185, - "pct_cuda_time": 1.1183353329846286, + "cuda_time_us": 62.993, + "pct_cuda_time": 1.0922375759341187, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15237,9 +15237,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.576, - "cuda_time_us": 8.492, - "pct_cuda_time": 0.1479614185199886, + "cpu_time_us": 21.041, + "cuda_time_us": 8.172, + "pct_cuda_time": 0.14169456083268964, "trace": "" }, "children": [ @@ -15247,8 +15247,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.492, - "pct_cuda_time": 0.1479614185199886, + "cuda_time_us": 8.172, + "pct_cuda_time": 0.14169456083268964, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -15258,9 +15258,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.608, - "cuda_time_us": 37.275, - "pct_cuda_time": 0.6494656000156115, + "cpu_time_us": 50.094, + "cuda_time_us": 34.851, + "pct_cuda_time": 0.6042825672515988, "trace": "" }, "children": [ @@ -15268,8 +15268,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 37.275, - "pct_cuda_time": 0.6494656000156115, + "cuda_time_us": 34.851, + "pct_cuda_time": 0.6042825672515988, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -15283,18 +15283,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 481.148, - "cuda_time_us": 157.985, - "pct_cuda_time": 2.752671302976966, + "cpu_time_us": 530.983, + "cuda_time_us": 163.283, + "pct_cuda_time": 2.831168988796385, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.515, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cpu_time_us": 24.497, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -15302,8 +15302,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15313,18 +15313,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 286.778, - "cuda_time_us": 47.095000000000006, - "pct_cuda_time": 0.8205655917568137, + "cpu_time_us": 315.251, + "cuda_time_us": 50.269999999999996, + "pct_cuda_time": 0.8716330852984957, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 47.871, - "cuda_time_us": 16.754, - "pct_cuda_time": 0.29191540342485733, + "cpu_time_us": 47.241, + "cuda_time_us": 16.914, + "pct_cuda_time": 0.29327236930055217, "trace": "" }, "children": [ @@ -15332,8 +15332,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.754, - "pct_cuda_time": 0.29191540342485733, + "cuda_time_us": 16.914, + "pct_cuda_time": 0.29327236930055217, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15343,9 +15343,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 23.095, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cpu_time_us": 27.751, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "" }, "children": [ @@ -15353,8 +15353,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -15364,9 +15364,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 133.069, - "cuda_time_us": 12.576, - "pct_cuda_time": 0.21911950062498542, + "cpu_time_us": 152.818, + "cuda_time_us": 14.9, + "pct_cuda_time": 0.2583515609896079, "trace": "" }, "children": [ @@ -15374,8 +15374,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.883, - "pct_cuda_time": 0.05023230918430605, + "cuda_time_us": 2.884, + "pct_cuda_time": 0.050005765227787186, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15384,8 +15384,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.688, - "pct_cuda_time": 0.09910557566435409, + "cuda_time_us": 8.252, + "pct_cuda_time": 0.14308168330780163, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15394,8 +15394,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.005, - "pct_cuda_time": 0.06978161577632526, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15405,9 +15405,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.343, - "cuda_time_us": 13.88, - "pct_cuda_time": 0.24183990686027337, + "cpu_time_us": 55.152, + "cuda_time_us": 14.611, + "pct_cuda_time": 0.2533405810482658, "trace": "" }, "children": [ @@ -15415,8 +15415,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 13.88, - "pct_cuda_time": 0.24183990686027337, + "cuda_time_us": 14.611, + "pct_cuda_time": 0.2533405810482658, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15428,9 +15428,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 27.331, - "cuda_time_us": 5.287, - "pct_cuda_time": 0.09211870227451478, + "cpu_time_us": 27.732, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -15438,8 +15438,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.287, - "pct_cuda_time": 0.09211870227451478, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15449,18 +15449,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 127.02, - "cuda_time_us": 100.958, - "pct_cuda_time": 1.7590542735446308, + "cpu_time_us": 146.009, + "cuda_time_us": 103.521, + "pct_cuda_time": 1.794953821825852, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 48.262, - "cuda_time_us": 61.301, - "pct_cuda_time": 1.068085600175909, + "cpu_time_us": 55.964, + "cuda_time_us": 61.811, + "pct_cuda_time": 1.071742841364339, "trace": "" }, "children": [ @@ -15468,8 +15468,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.301, - "pct_cuda_time": 1.068085600175909, + "cuda_time_us": 61.811, + "pct_cuda_time": 1.071742841364339, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15479,9 +15479,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.937, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cpu_time_us": 20.871, + "cuda_time_us": 8.171, + "pct_cuda_time": 0.14167722180175069, "trace": "" }, "children": [ @@ -15489,8 +15489,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cuda_time_us": 8.171, + "pct_cuda_time": 0.14167722180175069, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -15500,9 +15500,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.969, - "cuda_time_us": 31.526, - "pct_cuda_time": 0.5492971832620301, + "cpu_time_us": 51.446, + "cuda_time_us": 33.539, + "pct_cuda_time": 0.5815337586597623, "trace": "" }, "children": [ @@ -15510,8 +15510,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 31.526, - "pct_cuda_time": 0.5492971832620301, + "cuda_time_us": 33.539, + "pct_cuda_time": 0.5815337586597623, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -15525,18 +15525,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 602.8, - "cuda_time_us": 162.351, - "pct_cuda_time": 2.8287428471665876, + "cpu_time_us": 650.032, + "cuda_time_us": 164.68400000000003, + "pct_cuda_time": 2.855460971141784, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.035, - "cuda_time_us": 4.565, - "pct_cuda_time": 0.07953884544792132, + "cpu_time_us": 26.559, + "cuda_time_us": 4.846, + "pct_cuda_time": 0.08402494392990871, "trace": "" }, "children": [ @@ -15544,8 +15544,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.565, - "pct_cuda_time": 0.07953884544792132, + "cuda_time_us": 4.846, + "pct_cuda_time": 0.08402494392990871, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15555,18 +15555,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 308.791, - "cuda_time_us": 47.63600000000001, - "pct_cuda_time": 0.8299917725645521, + "cpu_time_us": 368.501, + "cuda_time_us": 49.688, + "pct_cuda_time": 0.8615417692920562, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 45.849, - "cuda_time_us": 16.734, - "pct_cuda_time": 0.2915669309365861, + "cpu_time_us": 51.046, + "cuda_time_us": 18.056, + "pct_cuda_time": 0.3130735426327758, "trace": "" }, "children": [ @@ -15574,8 +15574,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.734, - "pct_cuda_time": 0.2915669309365861, + "cuda_time_us": 18.056, + "pct_cuda_time": 0.3130735426327758, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15585,9 +15585,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.614, - "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "cpu_time_us": 23.144, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "" }, "children": [ @@ -15595,8 +15595,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.805, - "pct_cuda_time": 0.06629689089361239, + "cuda_time_us": 3.845, + "pct_cuda_time": 0.06666857396006995, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -15606,9 +15606,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 158.687, - "cuda_time_us": 12.736, - "pct_cuda_time": 0.22190728053115574, + "cpu_time_us": 197.515, + "cuda_time_us": 14.418, + "pct_cuda_time": 0.2499941480770581, "trace": "" }, "children": [ @@ -15616,8 +15616,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.603, - "pct_cuda_time": 0.045353694348508035, + "cuda_time_us": 2.643, + "pct_cuda_time": 0.045827058771512316, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15626,8 +15626,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.048, - "pct_cuda_time": 0.10537808045323727, + "cuda_time_us": 7.89, + "pct_cuda_time": 0.13680495410791985, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15636,8 +15636,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.085, - "pct_cuda_time": 0.07117550572941042, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15647,9 +15647,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.162, - "cuda_time_us": 14.361, - "pct_cuda_time": 0.2502206702031978, + "cpu_time_us": 61.993, + "cuda_time_us": 13.369, + "pct_cuda_time": 0.23180550462215216, "trace": "" }, "children": [ @@ -15657,8 +15657,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.361, - "pct_cuda_time": 0.2502206702031978, + "cuda_time_us": 13.369, + "pct_cuda_time": 0.23180550462215216, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15670,9 +15670,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 109.714, - "cuda_time_us": 5.047, - "pct_cuda_time": 0.08793703241525933, + "cpu_time_us": 91.717, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "" }, "children": [ @@ -15680,8 +15680,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.047, - "pct_cuda_time": 0.08793703241525933, + "cuda_time_us": 4.606, + "pct_cuda_time": 0.07986357650457274, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15691,18 +15691,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 143.344, - "cuda_time_us": 105.10300000000001, - "pct_cuda_time": 1.8312751967388554, + "cpu_time_us": 144.536, + "cuda_time_us": 105.54400000000001, + "pct_cuda_time": 1.8300306814152467, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 61.913, - "cuda_time_us": 63.804, - "pct_cuda_time": 1.1116969320830605, + "cpu_time_us": 58.928, + "cuda_time_us": 61.941, + "pct_cuda_time": 1.073996915386396, "trace": "" }, "children": [ @@ -15710,8 +15710,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.804, - "pct_cuda_time": 1.1116969320830605, + "cuda_time_us": 61.941, + "pct_cuda_time": 1.073996915386396, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15721,9 +15721,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 18.618, - "cuda_time_us": 8.051, - "pct_cuda_time": 0.14027760015360669, + "cpu_time_us": 18.768, + "cuda_time_us": 8.131, + "pct_cuda_time": 0.14098366056419473, "trace": "" }, "children": [ @@ -15731,8 +15731,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.051, - "pct_cuda_time": 0.14027760015360669, + "cuda_time_us": 8.131, + "pct_cuda_time": 0.14098366056419473, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -15742,9 +15742,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.88, - "cuda_time_us": 33.248, - "pct_cuda_time": 0.5793006645021879, + "cpu_time_us": 47.421, + "cuda_time_us": 35.472, + "pct_cuda_time": 0.6150501054646558, "trace": "" }, "children": [ @@ -15752,8 +15752,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 33.248, - "pct_cuda_time": 0.5793006645021879, + "cuda_time_us": 35.472, + "pct_cuda_time": 0.6150501054646558, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -15767,18 +15767,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 492.195, - "cuda_time_us": 161.511, - "pct_cuda_time": 2.8141070026591932, + "cpu_time_us": 549.622, + "cuda_time_us": 163.654, + "pct_cuda_time": 2.837601769274717, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 29.384, - "cuda_time_us": 4.767, - "pct_cuda_time": 0.08305841757946132, + "cpu_time_us": 28.362, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -15786,8 +15786,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.767, - "pct_cuda_time": 0.08305841757946132, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15797,18 +15797,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 297.023, - "cuda_time_us": 47.49499999999999, - "pct_cuda_time": 0.8275350415222392, + "cpu_time_us": 336.352, + "cuda_time_us": 47.665, + "pct_cuda_time": 0.8264649097026616, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 51.817, - "cuda_time_us": 17.826, - "pct_cuda_time": 0.3105935287961983, + "cpu_time_us": 52.789, + "cuda_time_us": 16.854, + "pct_cuda_time": 0.29223202744421817, "trace": "" }, "children": [ @@ -15816,8 +15816,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.826, - "pct_cuda_time": 0.3105935287961983, + "cuda_time_us": 16.854, + "pct_cuda_time": 0.29223202744421817, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15827,9 +15827,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.924, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cpu_time_us": 23.565, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "" }, "children": [ @@ -15837,8 +15837,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -15848,9 +15848,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 138.978, - "cuda_time_us": 12.815999999999999, - "pct_cuda_time": 0.22330117048424084, + "cpu_time_us": 170.164, + "cuda_time_us": 13.577, + "pct_cuda_time": 0.23541202305744335, "trace": "" }, "children": [ @@ -15858,8 +15858,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.643, - "pct_cuda_time": 0.04605063932505061, + "cuda_time_us": 2.523, + "pct_cuda_time": 0.043746375058844336, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15868,8 +15868,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.168, - "pct_cuda_time": 0.10746891538286499, + "cuda_time_us": 7.29, + "pct_cuda_time": 0.12640153554457995, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15878,8 +15878,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.005, - "pct_cuda_time": 0.06978161577632526, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -15889,9 +15889,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.322, - "cuda_time_us": 13.008, - "pct_cuda_time": 0.22664650637164518, + "cpu_time_us": 58.407, + "cuda_time_us": 13.469, + "pct_cuda_time": 0.23353940771604215, "trace": "" }, "children": [ @@ -15899,8 +15899,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 13.008, - "pct_cuda_time": 0.22664650637164518, + "cuda_time_us": 13.469, + "pct_cuda_time": 0.23353940771604215, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15912,9 +15912,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.238, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.05, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -15922,8 +15922,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -15933,18 +15933,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 123.013, - "cuda_time_us": 104.68299999999999, - "pct_cuda_time": 1.8239572744851578, + "cpu_time_us": 136.605, + "cuda_time_us": 106.697, + "pct_cuda_time": 1.850022584087798, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 45.788, - "cuda_time_us": 61.481, - "pct_cuda_time": 1.0712218525703505, + "cpu_time_us": 53.16, + "cuda_time_us": 63.404, + "pct_cuda_time": 1.0993639176500065, "trace": "" }, "children": [ @@ -15952,8 +15952,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 61.481, - "pct_cuda_time": 1.0712218525703505, + "cuda_time_us": 63.404, + "pct_cuda_time": 1.0993639176500065, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -15963,9 +15963,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.366, - "cuda_time_us": 8.772, - "pct_cuda_time": 0.15284003335578658, + "cpu_time_us": 18.578, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "" }, "children": [ @@ -15973,8 +15973,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.772, - "pct_cuda_time": 0.15284003335578658, + "cuda_time_us": 8.211, + "pct_cuda_time": 0.14237078303930673, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -15984,9 +15984,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 45.387, - "cuda_time_us": 34.43, - "pct_cuda_time": 0.599895388559021, + "cpu_time_us": 49.434, + "cuda_time_us": 35.082, + "pct_cuda_time": 0.6082878833984847, "trace": "" }, "children": [ @@ -15994,8 +15994,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 34.43, - "pct_cuda_time": 0.599895388559021, + "cuda_time_us": 35.082, + "pct_cuda_time": 0.6082878833984847, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -16009,18 +16009,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 467.427, - "cuda_time_us": 167.881, - "pct_cuda_time": 2.925095490173599, + "cpu_time_us": 531.204, + "cuda_time_us": 166.76, + "pct_cuda_time": 2.89145679937094, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.655, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 24.386, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -16028,8 +16028,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16039,18 +16039,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 279.127, - "cuda_time_us": 48.678, - "pct_cuda_time": 0.848147189203486, + "cpu_time_us": 319.598, + "cuda_time_us": 50.781, + "pct_cuda_time": 0.8804933301082736, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 44.917, - "cuda_time_us": 17.896, - "pct_cuda_time": 0.31181318250514783, + "cpu_time_us": 47.532, + "cuda_time_us": 17.776, + "pct_cuda_time": 0.3082186139698838, "trace": "" }, "children": [ @@ -16058,8 +16058,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.896, - "pct_cuda_time": 0.31181318250514783, + "cuda_time_us": 17.776, + "pct_cuda_time": 0.3082186139698838, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16069,9 +16069,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.442, + "cpu_time_us": 25.799, "cuda_time_us": 3.765, - "pct_cuda_time": 0.06559994591706982, + "pct_cuda_time": 0.06528145148495795, "trace": "" }, "children": [ @@ -16080,7 +16080,7 @@ "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, "cuda_time_us": 3.765, - "pct_cuda_time": 0.06559994591706982, + "pct_cuda_time": 0.06528145148495795, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -16090,9 +16090,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 132.578, - "cuda_time_us": 12.897, - "pct_cuda_time": 0.22471248406173958, + "cpu_time_us": 155.984, + "cuda_time_us": 14.297999999999998, + "pct_cuda_time": 0.24791346436439013, "trace": "" }, "children": [ @@ -16100,8 +16100,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.563, - "pct_cuda_time": 0.04465674937196546, + "cuda_time_us": 2.523, + "pct_cuda_time": 0.043746375058844336, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16110,8 +16110,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.369, - "pct_cuda_time": 0.11097106388999142, + "cuda_time_us": 7.85, + "pct_cuda_time": 0.13611139287036386, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16120,8 +16120,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.925, + "pct_cuda_time": 0.06805569643518193, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16131,9 +16131,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 53.28, - "cuda_time_us": 14.12, - "pct_cuda_time": 0.24602157671952876, + "cpu_time_us": 59.859, + "cuda_time_us": 14.942, + "pct_cuda_time": 0.25907980028904165, "trace": "" }, "children": [ @@ -16141,8 +16141,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.12, - "pct_cuda_time": 0.24602157671952876, + "cuda_time_us": 14.942, + "pct_cuda_time": 0.25907980028904165, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16154,9 +16154,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.287, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cpu_time_us": 29.274, + "cuda_time_us": 5.046, + "pct_cuda_time": 0.08749275011768867, "trace": "" }, "children": [ @@ -16164,8 +16164,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cuda_time_us": 5.046, + "pct_cuda_time": 0.08749275011768867, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16175,18 +16175,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 123.975, - "cuda_time_us": 109.871, - "pct_cuda_time": 1.9143510379427298, + "cpu_time_us": 140.83, + "cuda_time_us": 106.287, + "pct_cuda_time": 1.842913581402849, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 47.591, - "cuda_time_us": 65.326, - "pct_cuda_time": 1.1382156884405052, + "cpu_time_us": 51.737, + "cuda_time_us": 63.064, + "pct_cuda_time": 1.0934686471307804, "trace": "" }, "children": [ @@ -16194,8 +16194,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 65.326, - "pct_cuda_time": 1.1382156884405052, + "cuda_time_us": 63.064, + "pct_cuda_time": 1.0934686471307804, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16205,9 +16205,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.566, - "cuda_time_us": 8.452, - "pct_cuda_time": 0.147264473543446, + "cpu_time_us": 20.761, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "" }, "children": [ @@ -16215,8 +16215,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.452, - "pct_cuda_time": 0.147264473543446, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -16226,9 +16226,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 43.966, - "cuda_time_us": 36.093, - "pct_cuda_time": 0.6288708759587786, + "cpu_time_us": 51.316, + "cuda_time_us": 34.932, + "pct_cuda_time": 0.6056870287576498, "trace": "" }, "children": [ @@ -16236,8 +16236,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 36.093, - "pct_cuda_time": 0.6288708759587786, + "cuda_time_us": 34.932, + "pct_cuda_time": 0.6056870287576498, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -16251,18 +16251,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 509.521, - "cuda_time_us": 162.221, - "pct_cuda_time": 2.8264777759928243, + "cpu_time_us": 646.136, + "cuda_time_us": 170.093, + "pct_cuda_time": 2.949247789490293, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.084, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.08304099395504776, + "cpu_time_us": 25.078, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -16270,8 +16270,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.766, - "pct_cuda_time": 0.08304099395504776, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16281,18 +16281,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 300.218, - "cuda_time_us": 48.997, - "pct_cuda_time": 0.8537053253914131, + "cpu_time_us": 413.177, + "cuda_time_us": 52.713, + "pct_cuda_time": 0.913992337882228, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 44.326, - "cuda_time_us": 17.715, - "pct_cuda_time": 0.3086595064862927, + "cpu_time_us": 50.936, + "cuda_time_us": 17.195, + "pct_cuda_time": 0.29814463699438304, "trace": "" }, "children": [ @@ -16300,8 +16300,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.715, - "pct_cuda_time": 0.3086595064862927, + "cuda_time_us": 17.195, + "pct_cuda_time": 0.29814463699438304, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16311,9 +16311,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 28.953, - "cuda_time_us": 3.765, - "pct_cuda_time": 0.06559994591706982, + "cpu_time_us": 37.566, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "" }, "children": [ @@ -16321,8 +16321,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.765, - "pct_cuda_time": 0.06559994591706982, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -16332,9 +16332,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 133.48, - "cuda_time_us": 13.657, - "pct_cuda_time": 0.2379544386160485, + "cpu_time_us": 221.311, + "cuda_time_us": 17.063, + "pct_cuda_time": 0.29585588491044823, "trace": "" }, "children": [ @@ -16342,8 +16342,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.643, - "pct_cuda_time": 0.04605063932505061, + "cuda_time_us": 5.688, + "pct_cuda_time": 0.09862440798046238, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16352,8 +16352,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.969, - "pct_cuda_time": 0.12142523853813006, + "cuda_time_us": 7.691, + "pct_cuda_time": 0.1333544869510788, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16362,8 +16362,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.045, - "pct_cuda_time": 0.07047856075286783, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16373,9 +16373,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 56.494, - "cuda_time_us": 13.86, - "pct_cuda_time": 0.24149143437200205, + "cpu_time_us": 69.224, + "cuda_time_us": 14.691, + "pct_cuda_time": 0.25472770352337776, "trace": "" }, "children": [ @@ -16383,8 +16383,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 13.86, - "pct_cuda_time": 0.24149143437200205, + "cuda_time_us": 14.691, + "pct_cuda_time": 0.25472770352337776, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16396,9 +16396,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 28.312, + "cpu_time_us": 31.577, "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -16407,7 +16407,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 4.726, - "pct_cuda_time": 0.08234404897850517, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16417,18 +16417,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 141.731, - "cuda_time_us": 103.732, - "pct_cuda_time": 1.8073874076678584, + "cpu_time_us": 158.447, + "cuda_time_us": 107.96799999999999, + "pct_cuda_time": 1.8720604924111393, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 61.702, - "cuda_time_us": 62.693, - "pct_cuda_time": 1.0923392853595904, + "cpu_time_us": 66.329, + "cuda_time_us": 64.095, + "pct_cuda_time": 1.1113451880287861, "trace": "" }, "children": [ @@ -16436,8 +16436,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.693, - "pct_cuda_time": 1.0923392853595904, + "cuda_time_us": 64.095, + "pct_cuda_time": 1.1113451880287861, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16447,9 +16447,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.706, - "cuda_time_us": 8.171, - "pct_cuda_time": 0.14236843508323438, + "cpu_time_us": 21.512, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "" }, "children": [ @@ -16457,8 +16457,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.171, - "pct_cuda_time": 0.14236843508323438, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -16468,9 +16468,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.81, - "cuda_time_us": 32.868, - "pct_cuda_time": 0.5726796872250335, + "cpu_time_us": 52.929, + "cuda_time_us": 35.582, + "pct_cuda_time": 0.6169573988679347, "trace": "" }, "children": [ @@ -16478,8 +16478,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 32.868, - "pct_cuda_time": 0.5726796872250335, + "cuda_time_us": 35.582, + "pct_cuda_time": 0.6169573988679347, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -16493,18 +16493,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 504.664, - "cuda_time_us": 165.92700000000002, - "pct_cuda_time": 2.8910497280694942, + "cpu_time_us": 566.737, + "cuda_time_us": 165.908, + "pct_cuda_time": 2.876683945010997, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.677, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cpu_time_us": 27.652, + "cuda_time_us": 4.806, + "pct_cuda_time": 0.08333138269235271, "trace": "" }, "children": [ @@ -16512,8 +16512,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cuda_time_us": 4.806, + "pct_cuda_time": 0.08333138269235271, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16523,18 +16523,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 298.125, - "cuda_time_us": 48.908, - "pct_cuda_time": 0.8521546228186059, + "cpu_time_us": 343.554, + "cuda_time_us": 50.921, + "pct_cuda_time": 0.8829207944397196, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.279, - "cuda_time_us": 17.265, - "pct_cuda_time": 0.3008188755001887, + "cpu_time_us": 50.645, + "cuda_time_us": 17.966, + "pct_cuda_time": 0.31151302984827484, "trace": "" }, "children": [ @@ -16542,8 +16542,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.265, - "pct_cuda_time": 0.3008188755001887, + "cuda_time_us": 17.966, + "pct_cuda_time": 0.31151302984827484, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16553,9 +16553,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 23.184, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cpu_time_us": 25.638, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "" }, "children": [ @@ -16563,8 +16563,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -16574,9 +16574,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 135.693, - "cuda_time_us": 13.017, - "pct_cuda_time": 0.2268033189913673, + "cpu_time_us": 167.731, + "cuda_time_us": 15.58, + "pct_cuda_time": 0.2701421020280597, "trace": "" }, "children": [ @@ -16584,8 +16584,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.763, - "pct_cuda_time": 0.04814147425467833, + "cuda_time_us": 2.883, + "pct_cuda_time": 0.04998842619684828, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16594,8 +16594,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.249, - "pct_cuda_time": 0.10888022896036369, + "cuda_time_us": 8.812, + "pct_cuda_time": 0.15279154063358552, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16604,8 +16604,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.005, - "pct_cuda_time": 0.06978161577632526, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16615,9 +16615,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 65.328, - "cuda_time_us": 14.661, - "pct_cuda_time": 0.2554477575272671, + "cpu_time_us": 66.259, + "cuda_time_us": 13.49, + "pct_cuda_time": 0.23390352736575906, "trace": "" }, "children": [ @@ -16625,8 +16625,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.661, - "pct_cuda_time": 0.2554477575272671, + "cuda_time_us": 13.49, + "pct_cuda_time": 0.23390352736575906, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16638,9 +16638,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.98, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cpu_time_us": 28.102, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -16648,8 +16648,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.806, - "pct_cuda_time": 0.08373793893159033, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16659,18 +16659,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 132.878, - "cuda_time_us": 107.56700000000001, - "pct_cuda_time": 1.874207007293878, + "cpu_time_us": 149.383, + "cuda_time_us": 105.455, + "pct_cuda_time": 1.8284875076616844, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 54.261, - "cuda_time_us": 62.903, - "pct_cuda_time": 1.095998246486439, + "cpu_time_us": 58.688, + "cuda_time_us": 63.404, + "pct_cuda_time": 1.0993639176500065, "trace": "" }, "children": [ @@ -16678,8 +16678,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.903, - "pct_cuda_time": 1.095998246486439, + "cuda_time_us": 63.404, + "pct_cuda_time": 1.0993639176500065, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16689,9 +16689,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.546, - "cuda_time_us": 9.693, - "pct_cuda_time": 0.16888719144067935, + "cpu_time_us": 22.113, + "cuda_time_us": 8.051, + "pct_cuda_time": 0.13959653808908273, "trace": "" }, "children": [ @@ -16699,8 +16699,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 9.693, - "pct_cuda_time": 0.16888719144067935, + "cuda_time_us": 8.051, + "pct_cuda_time": 0.13959653808908273, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -16710,9 +16710,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.119, - "cuda_time_us": 34.971, - "pct_cuda_time": 0.6093215693667592, + "cpu_time_us": 51.317, + "cuda_time_us": 34.0, + "pct_cuda_time": 0.589527051922595, "trace": "" }, "children": [ @@ -16720,8 +16720,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 34.971, - "pct_cuda_time": 0.6093215693667592, + "cuda_time_us": 34.0, + "pct_cuda_time": 0.589527051922595, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -16735,18 +16735,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 491.043, - "cuda_time_us": 161.49, - "pct_cuda_time": 2.8137411065465088, + "cpu_time_us": 576.312, + "cuda_time_us": 166.14600000000002, + "pct_cuda_time": 2.8808106343744555, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.276, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 27.832, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "" }, "children": [ @@ -16754,8 +16754,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16765,18 +16765,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 291.174, - "cuda_time_us": 47.595, - "pct_cuda_time": 0.8292774039635957, + "cpu_time_us": 351.386, + "cuda_time_us": 50.689, + "pct_cuda_time": 0.8788981392618948, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 44.947, - "cuda_time_us": 16.864, - "pct_cuda_time": 0.29383200211034943, + "cpu_time_us": 57.165, + "cuda_time_us": 16.915, + "pct_cuda_time": 0.29328970833149104, "trace": "" }, "children": [ @@ -16784,8 +16784,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 16.864, - "pct_cuda_time": 0.29383200211034943, + "cuda_time_us": 16.915, + "pct_cuda_time": 0.29328970833149104, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16795,9 +16795,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.883, - "cuda_time_us": 3.725, - "pct_cuda_time": 0.06490300094052724, + "cpu_time_us": 25.268, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "" }, "children": [ @@ -16805,8 +16805,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.725, - "pct_cuda_time": 0.06490300094052724, + "cuda_time_us": 3.764, + "pct_cuda_time": 0.06526411245401906, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -16816,9 +16816,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 141.191, - "cuda_time_us": 12.776, - "pct_cuda_time": 0.22260422550769826, + "cpu_time_us": 174.701, + "cuda_time_us": 16.02, + "pct_cuda_time": 0.2777712756411757, "trace": "" }, "children": [ @@ -16826,8 +16826,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.683, - "pct_cuda_time": 0.04674758430159318, + "cuda_time_us": 2.883, + "pct_cuda_time": 0.04998842619684828, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16836,8 +16836,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.128, - "pct_cuda_time": 0.10677197040632241, + "cuda_time_us": 9.333, + "pct_cuda_time": 0.16182517575275235, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16846,8 +16846,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.965, - "pct_cuda_time": 0.06908467079978269, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -16857,9 +16857,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.202, - "cuda_time_us": 14.23, - "pct_cuda_time": 0.24793817540502086, + "cpu_time_us": 62.533, + "cuda_time_us": 13.99, + "pct_cuda_time": 0.242573042835209, "trace": "" }, "children": [ @@ -16867,8 +16867,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.23, - "pct_cuda_time": 0.24793817540502086, + "cuda_time_us": 13.99, + "pct_cuda_time": 0.242573042835209, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16880,9 +16880,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.768, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cpu_time_us": 27.952, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "" }, "children": [ @@ -16890,8 +16890,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.686, - "pct_cuda_time": 0.0816471040019626, + "cuda_time_us": 4.766, + "pct_cuda_time": 0.0826378214547967, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -16901,18 +16901,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 134.52, - "cuda_time_us": 104.523, - "pct_cuda_time": 1.8211694945789878, + "cpu_time_us": 150.104, + "cuda_time_us": 105.965, + "pct_cuda_time": 1.8373304134405233, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 54.161, - "cuda_time_us": 62.162, - "pct_cuda_time": 1.0830873407959878, + "cpu_time_us": 61.993, + "cuda_time_us": 62.282, + "pct_cuda_time": 1.079909524936561, "trace": "" }, "children": [ @@ -16920,8 +16920,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.162, - "pct_cuda_time": 1.0830873407959878, + "cuda_time_us": 62.282, + "pct_cuda_time": 1.079909524936561, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -16931,9 +16931,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.747, - "cuda_time_us": 8.091, - "pct_cuda_time": 0.14097454513014926, + "cpu_time_us": 19.329, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "" }, "children": [ @@ -16941,8 +16941,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.091, - "pct_cuda_time": 0.14097454513014926, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -16952,9 +16952,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.91, - "cuda_time_us": 34.27, - "pct_cuda_time": 0.5971076086528507, + "cpu_time_us": 52.428, + "cuda_time_us": 35.392, + "pct_cuda_time": 0.6136629829895437, "trace": "" }, "children": [ @@ -16962,8 +16962,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 34.27, - "pct_cuda_time": 0.5971076086528507, + "cuda_time_us": 35.392, + "pct_cuda_time": 0.6136629829895437, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -16977,18 +16977,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 476.221, - "cuda_time_us": 165.978, - "pct_cuda_time": 2.891938332914586, + "cpu_time_us": 553.047, + "cuda_time_us": 163.282, + "pct_cuda_time": 2.8311516497654465, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.705, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.08025321404887745, + "cpu_time_us": 27.411, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -16996,8 +16996,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.606, - "pct_cuda_time": 0.08025321404887745, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17007,18 +17007,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 280.058, - "cuda_time_us": 48.577000000000005, - "pct_cuda_time": 0.8463874031377161, + "cpu_time_us": 329.372, + "cuda_time_us": 48.325, + "pct_cuda_time": 0.8379086701223356, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 45.288, - "cuda_time_us": 17.315, - "pct_cuda_time": 0.30169005672086696, + "cpu_time_us": 52.689, + "cuda_time_us": 16.574, + "pct_cuda_time": 0.2873770987813262, "trace": "" }, "children": [ @@ -17026,8 +17026,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.315, - "pct_cuda_time": 0.30169005672086696, + "cuda_time_us": 16.574, + "pct_cuda_time": 0.2873770987813262, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17037,9 +17037,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.103, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cpu_time_us": 23.745, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "" }, "children": [ @@ -17047,8 +17047,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.885, - "pct_cuda_time": 0.06769078084669755, + "cuda_time_us": 3.765, + "pct_cuda_time": 0.06528145148495795, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -17058,9 +17058,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 132.468, - "cuda_time_us": 13.017000000000001, - "pct_cuda_time": 0.22680331899136733, + "cpu_time_us": 163.194, + "cuda_time_us": 14.137, + "pct_cuda_time": 0.24512188038322727, "trace": "" }, "children": [ @@ -17068,8 +17068,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.483, - "pct_cuda_time": 0.04326285941888031, + "cuda_time_us": 2.562, + "pct_cuda_time": 0.044422597265461425, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17078,8 +17078,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.448, - "pct_cuda_time": 0.11234753021866302, + "cuda_time_us": 7.69, + "pct_cuda_time": 0.1333371479201399, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17088,8 +17088,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.086, - "pct_cuda_time": 0.07119292935382399, + "cuda_time_us": 3.885, + "pct_cuda_time": 0.06736213519762593, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17099,9 +17099,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 52.649, - "cuda_time_us": 14.36, - "pct_cuda_time": 0.25020324657878423, + "cpu_time_us": 56.164, + "cuda_time_us": 13.849, + "pct_cuda_time": 0.24012823947282413, "trace": "" }, "children": [ @@ -17109,8 +17109,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.36, - "pct_cuda_time": 0.25020324657878423, + "cuda_time_us": 13.849, + "pct_cuda_time": 0.24012823947282413, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17122,9 +17122,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.887, + "cpu_time_us": 30.095, "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -17133,7 +17133,7 @@ "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17143,18 +17143,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 132.598, - "cuda_time_us": 108.149, - "pct_cuda_time": 1.8843475567025723, + "cpu_time_us": 145.477, + "cuda_time_us": 105.625, + "pct_cuda_time": 1.8314351429212974, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 54.832, - "cuda_time_us": 63.624, - "pct_cuda_time": 1.108560679688619, + "cpu_time_us": 59.138, + "cuda_time_us": 63.694, + "pct_cuda_time": 1.1043922366222874, "trace": "" }, "children": [ @@ -17162,8 +17162,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.624, - "pct_cuda_time": 1.108560679688619, + "cuda_time_us": 63.694, + "pct_cuda_time": 1.1043922366222874, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17173,9 +17173,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.136, - "cuda_time_us": 8.292, - "pct_cuda_time": 0.14447669363727567, + "cpu_time_us": 20.411, + "cuda_time_us": 8.251, + "pct_cuda_time": 0.1430643442768627, "trace": "" }, "children": [ @@ -17183,8 +17183,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.292, - "pct_cuda_time": 0.14447669363727567, + "cuda_time_us": 8.251, + "pct_cuda_time": 0.1430643442768627, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -17194,9 +17194,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 44.897, - "cuda_time_us": 36.233, - "pct_cuda_time": 0.6313101833766775, + "cpu_time_us": 49.514, + "cuda_time_us": 33.68, + "pct_cuda_time": 0.5839785620221472, "trace": "" }, "children": [ @@ -17204,8 +17204,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 36.233, - "pct_cuda_time": 0.6313101833766775, + "cuda_time_us": 33.68, + "pct_cuda_time": 0.5839785620221472, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -17219,18 +17219,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 501.038, - "cuda_time_us": 163.602, - "pct_cuda_time": 2.850539801307957, + "cpu_time_us": 561.72, + "cuda_time_us": 163.793, + "pct_cuda_time": 2.840011894575224, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 23.655, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.531, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -17238,8 +17238,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17249,18 +17249,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 281.08, - "cuda_time_us": 49.909, - "pct_cuda_time": 0.8695956708565837, + "cpu_time_us": 333.779, + "cuda_time_us": 47.414, + "pct_cuda_time": 0.8221128129369978, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 44.467, - "cuda_time_us": 18.436, - "pct_cuda_time": 0.32122193968847257, + "cpu_time_us": 50.235, + "cuda_time_us": 17.615, + "pct_cuda_time": 0.3054270299887209, "trace": "" }, "children": [ @@ -17268,8 +17268,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 18.436, - "pct_cuda_time": 0.32122193968847257, + "cuda_time_us": 17.615, + "pct_cuda_time": 0.3054270299887209, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17279,9 +17279,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.002, - "cuda_time_us": 3.685, - "pct_cuda_time": 0.06420605596398468, + "cpu_time_us": 24.266, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "" }, "children": [ @@ -17289,8 +17289,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.685, - "pct_cuda_time": 0.06420605596398468, + "cuda_time_us": 3.684, + "pct_cuda_time": 0.06387698997890708, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -17300,9 +17300,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 134.501, - "cuda_time_us": 13.578, - "pct_cuda_time": 0.23657797228737687, + "cpu_time_us": 166.889, + "cuda_time_us": 13.217, + "pct_cuda_time": 0.2291699719194394, "trace": "" }, "children": [ @@ -17310,8 +17310,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.643, - "pct_cuda_time": 0.04605063932505061, + "cuda_time_us": 2.603, + "pct_cuda_time": 0.04513349753395633, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17320,8 +17320,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 7.01, - "pct_cuda_time": 0.12213960713908618, + "cuda_time_us": 6.889, + "pct_cuda_time": 0.11944858413808111, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17330,8 +17330,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 3.925, - "pct_cuda_time": 0.06838772582324011, + "cuda_time_us": 3.725, + "pct_cuda_time": 0.06458789024740197, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17341,9 +17341,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 53.319, - "cuda_time_us": 14.21, - "pct_cuda_time": 0.2475897029167496, + "cpu_time_us": 59.539, + "cuda_time_us": 12.898, + "pct_cuda_time": 0.22363882104993035, "trace": "" }, "children": [ @@ -17351,8 +17351,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.21, - "pct_cuda_time": 0.2475897029167496, + "cuda_time_us": 12.898, + "pct_cuda_time": 0.22363882104993035, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17364,9 +17364,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.369, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cpu_time_us": 29.004, + "cuda_time_us": 4.886, + "pct_cuda_time": 0.0847185051674647, "trace": "" }, "children": [ @@ -17374,8 +17374,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.645, - "pct_cuda_time": 0.08093273540100646, + "cuda_time_us": 4.886, + "pct_cuda_time": 0.0847185051674647, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17385,18 +17385,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 147.89, - "cuda_time_us": 104.482, - "pct_cuda_time": 1.8204551259780315, + "cpu_time_us": 149.453, + "cuda_time_us": 106.84700000000001, + "pct_cuda_time": 1.852623438728633, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 68.192, - "cuda_time_us": 62.382, - "pct_cuda_time": 1.086920538166972, + "cpu_time_us": 57.526, + "cuda_time_us": 63.865, + "pct_cuda_time": 1.1073572109128393, "trace": "" }, "children": [ @@ -17404,8 +17404,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 62.382, - "pct_cuda_time": 1.086920538166972, + "cuda_time_us": 63.865, + "pct_cuda_time": 1.1073572109128393, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17415,9 +17415,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 18.358, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cpu_time_us": 21.402, + "cuda_time_us": 8.331, + "pct_cuda_time": 0.1444514667519747, "trace": "" }, "children": [ @@ -17425,8 +17425,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.131, - "pct_cuda_time": 0.14167149010669183, + "cuda_time_us": 8.331, + "pct_cuda_time": 0.1444514667519747, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -17436,9 +17436,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.529, - "cuda_time_us": 33.969, - "pct_cuda_time": 0.5918630977043678, + "cpu_time_us": 52.979, + "cuda_time_us": 34.651, + "pct_cuda_time": 0.6008147610638189, "trace": "" }, "children": [ @@ -17446,8 +17446,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 33.969, - "pct_cuda_time": 0.5918630977043678, + "cuda_time_us": 34.651, + "pct_cuda_time": 0.6008147610638189, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -17461,18 +17461,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 504.343, - "cuda_time_us": 161.89999999999998, - "pct_cuda_time": 2.8208847925560696, + "cpu_time_us": 563.192, + "cuda_time_us": 164.404, + "pct_cuda_time": 2.850606042478892, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 31.226, - "cuda_time_us": 4.525, - "pct_cuda_time": 0.07884190047137875, + "cpu_time_us": 29.574, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "" }, "children": [ @@ -17480,8 +17480,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.525, - "pct_cuda_time": 0.07884190047137875, + "cuda_time_us": 4.646, + "pct_cuda_time": 0.08055713774212873, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17491,18 +17491,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 297.524, - "cuda_time_us": 48.675999999999995, - "pct_cuda_time": 0.8481123419546589, + "cpu_time_us": 335.101, + "cuda_time_us": 48.165, + "pct_cuda_time": 0.8351344251721114, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 50.546, - "cuda_time_us": 17.435, - "pct_cuda_time": 0.3037808916504946, + "cpu_time_us": 51.868, + "cuda_time_us": 16.884, + "pct_cuda_time": 0.2927521983723852, "trace": "" }, "children": [ @@ -17510,8 +17510,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 17.435, - "pct_cuda_time": 0.3037808916504946, + "cuda_time_us": 16.884, + "pct_cuda_time": 0.2927521983723852, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17521,9 +17521,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 22.143, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cpu_time_us": 25.057, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "" }, "children": [ @@ -17531,8 +17531,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cuda_time_us": 3.804, + "pct_cuda_time": 0.06595767369157504, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -17542,9 +17542,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 142.073, - "cuda_time_us": 12.936, - "pct_cuda_time": 0.22539200541386856, + "cpu_time_us": 166.229, + "cuda_time_us": 14.097999999999999, + "pct_cuda_time": 0.24444565817661018, "trace": "" }, "children": [ @@ -17552,8 +17552,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.563, - "pct_cuda_time": 0.04465674937196546, + "cuda_time_us": 2.643, + "pct_cuda_time": 0.045827058771512316, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17562,8 +17562,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 6.288, - "pct_cuda_time": 0.10955975031249271, + "cuda_time_us": 7.45, + "pct_cuda_time": 0.12917578049480394, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17572,8 +17572,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.085, - "pct_cuda_time": 0.07117550572941042, + "cuda_time_us": 4.005, + "pct_cuda_time": 0.06944281891029393, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17583,9 +17583,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 55.543, - "cuda_time_us": 14.46, - "pct_cuda_time": 0.2519456090201407, + "cpu_time_us": 56.995, + "cuda_time_us": 13.379, + "pct_cuda_time": 0.23197889493154117, "trace": "" }, "children": [ @@ -17593,8 +17593,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 14.46, - "pct_cuda_time": 0.2519456090201407, + "cuda_time_us": 13.379, + "pct_cuda_time": 0.23197889493154117, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17606,9 +17606,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 25.598, - "cuda_time_us": 5.047, - "pct_cuda_time": 0.08793703241525933, + "cpu_time_us": 27.551, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -17616,8 +17616,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 5.047, - "pct_cuda_time": 0.08793703241525933, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17627,18 +17627,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 134.181, - "cuda_time_us": 103.652, - "pct_cuda_time": 1.8059935177147732, + "cpu_time_us": 152.989, + "cuda_time_us": 106.90700000000001, + "pct_cuda_time": 1.853663780584967, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 55.513, - "cuda_time_us": 63.074, - "pct_cuda_time": 1.0989776862611587, + "cpu_time_us": 58.237, + "cuda_time_us": 64.275, + "pct_cuda_time": 1.1144662135977883, "trace": "" }, "children": [ @@ -17646,8 +17646,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.074, - "pct_cuda_time": 1.0989776862611587, + "cuda_time_us": 64.275, + "pct_cuda_time": 1.1144662135977883, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17657,9 +17657,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 17.396, - "cuda_time_us": 8.171, - "pct_cuda_time": 0.14236843508323438, + "cpu_time_us": 22.634, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "" }, "children": [ @@ -17667,8 +17667,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.171, - "pct_cuda_time": 0.14236843508323438, + "cuda_time_us": 8.291, + "pct_cuda_time": 0.14375790551441872, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -17678,9 +17678,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 46.009, - "cuda_time_us": 32.407, - "pct_cuda_time": 0.5646473963703802, + "cpu_time_us": 52.809, + "cuda_time_us": 34.341, + "pct_cuda_time": 0.59543966147276, "trace": "" }, "children": [ @@ -17688,8 +17688,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 32.407, - "pct_cuda_time": 0.5646473963703802, + "cuda_time_us": 34.341, + "pct_cuda_time": 0.59543966147276, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -17703,18 +17703,18 @@ { "entry": { "name": "LlamaDecoderLayer", - "cpu_time_us": 573.466, - "cuda_time_us": 149.702, - "pct_cuda_time": 2.6083514219594117, + "cpu_time_us": 663.192, + "cuda_time_us": 148.279, + "pct_cuda_time": 2.5710141685891315, "trace": "" }, "children": [ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.437, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cpu_time_us": 27.411, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "" }, "children": [ @@ -17722,8 +17722,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.566, - "pct_cuda_time": 0.07955626907233487, + "cuda_time_us": 4.686, + "pct_cuda_time": 0.08125069897968472, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17733,18 +17733,18 @@ { "entry": { "name": "LlamaAttention", - "cpu_time_us": 368.169, - "cuda_time_us": 42.827, - "pct_cuda_time": 0.7462015627597209, + "cpu_time_us": 432.537, + "cuda_time_us": 40.263, + "pct_cuda_time": 0.6981214026929248, "trace": "" }, "children": [ { "entry": { "name": "QKVParallelLinear(weight=bfloat16[6144, 4096])", - "cpu_time_us": 46.018, - "cuda_time_us": 15.832, - "pct_cuda_time": 0.275850821715551, + "cpu_time_us": 52.068, + "cuda_time_us": 14.07, + "pct_cuda_time": 0.24396016531032094, "trace": "" }, "children": [ @@ -17752,8 +17752,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 15.832, - "pct_cuda_time": 0.275850821715551, + "cuda_time_us": 14.07, + "pct_cuda_time": 0.24396016531032094, "trace": "_rocm_C::wvSplitK(bfloat16[6144, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17763,9 +17763,9 @@ { "entry": { "name": "Llama3RotaryEmbedding", - "cpu_time_us": 21.602, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cpu_time_us": 25.078, + "cuda_time_us": 3.884, + "pct_cuda_time": 0.06734479616668704, "trace": "" }, "children": [ @@ -17773,8 +17773,8 @@ "entry": { "name": "void vllm::rotary_embedding_kernel(long const*, c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, int, long, long, int, int, int)", "cpu_time_us": 0, - "cuda_time_us": 3.845, - "pct_cuda_time": 0.06699383587015498, + "cuda_time_us": 3.884, + "pct_cuda_time": 0.06734479616668704, "trace": "_C::rotary_embedding(int64[1], bfloat16[1, 4096], bfloat16[1, 1024], 128, bfloat16[131072, 128], True)" }, "children": [] @@ -17784,9 +17784,9 @@ { "entry": { "name": "Attention", - "cpu_time_us": 214.731, - "cuda_time_us": 12.495000000000001, - "pct_cuda_time": 0.2177081870474867, + "cpu_time_us": 254.641, + "cuda_time_us": 10.652999999999999, + "pct_cuda_time": 0.18471269659210013, "trace": "" }, "children": [ @@ -17794,8 +17794,8 @@ "entry": { "name": "void vllm::reshape_and_cache_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, __hip_bfloat16*, long const*, int, int, int, int, int, int, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 2.843, - "pct_cuda_time": 0.04953536420776348, + "cuda_time_us": 2.923, + "pct_cuda_time": 0.050681987434404274, "trace": "_C_cache_ops::reshape_and_cache(bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], int64[1], None, float32[], float32[]) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17804,8 +17804,8 @@ "entry": { "name": "void paged_attention_ll4mi_QKV_mfma4_kernel<__hip_bfloat16, __hip_bfloat16, (vllm::Fp8KVCacheDataType)0, __hip_bfloat16, 16, 128, 256, false, 4>(__hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16 const*, int, float, int const*, int const*, int const*, int, float const*, int, int, int, float*, float*, __hip_bfloat16*, __hip_bfloat16*, int, float const*, float const*, float const*)", "cpu_time_us": 0, - "cuda_time_us": 5.647, - "pct_cuda_time": 0.09839120706339796, + "cuda_time_us": 5.047, + "pct_cuda_time": 0.08751008914862757, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17814,8 +17814,8 @@ "entry": { "name": "void paged_attention_ll4mi_reduce_kernel<__hip_bfloat16, __hip_bfloat16, 128, 128, 256, 1>(__hip_bfloat16*, float const*, float const*, __hip_bfloat16 const*, int const*, int const*, int, float const*)", "cpu_time_us": 0, - "cuda_time_us": 4.005, - "pct_cuda_time": 0.06978161577632526, + "cuda_time_us": 2.683, + "pct_cuda_time": 0.04652062000906831, "trace": "_rocm_C::paged_attention(bfloat16[1, 32, 128], float32[1, 32, 2], float32[1, 32, 2], bfloat16[1, 32, 2, 128], bfloat16[1, 32, 128], bfloat16[79739, 8, 16, 16, 8], bfloat16[79739, 8, 128, 16], 8, 0.08838834764831845, int32[1, 17], int32[1], None, 16, 257, None, None, float32[], float32[], None) <- vllm::unified_attention_with_output(bfloat16[1, 32, 128], bfloat16[1, 8, 128], bfloat16[1, 8, 128], bfloat16[1, 32, 128], None, None)" }, "children": [] @@ -17825,9 +17825,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 4096])", - "cpu_time_us": 58.908, - "cuda_time_us": 10.655, - "pct_cuda_time": 0.18564871812652828, + "cpu_time_us": 64.256, + "cuda_time_us": 11.656, + "pct_cuda_time": 0.20210374462381672, "trace": "" }, "children": [ @@ -17835,8 +17835,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 10.655, - "pct_cuda_time": 0.18564871812652828, + "cuda_time_us": 11.656, + "pct_cuda_time": 0.20210374462381672, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17848,9 +17848,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 26.941, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cpu_time_us": 30.285, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.07847645402946074, "trace": "" }, "children": [ @@ -17858,8 +17858,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.646, - "pct_cuda_time": 0.08095015902542002, + "cuda_time_us": 4.526, + "pct_cuda_time": 0.07847645402946074, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17869,18 +17869,18 @@ { "entry": { "name": "LlamaMLP", - "cpu_time_us": 134.261, - "cuda_time_us": 97.66300000000001, - "pct_cuda_time": 1.7016434311019364, + "cpu_time_us": 152.969, + "cuda_time_us": 98.804, + "pct_cuda_time": 1.7131656128870614, "trace": "" }, "children": [ { "entry": { "name": "MergedColumnParallelLinear(weight=bfloat16[28672, 4096])", - "cpu_time_us": 54.301, - "cuda_time_us": 63.654, - "pct_cuda_time": 1.109083388421026, + "cpu_time_us": 61.252, + "cuda_time_us": 62.232, + "pct_cuda_time": 1.0790425733896158, "trace": "" }, "children": [ @@ -17888,8 +17888,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 63.654, - "pct_cuda_time": 1.109083388421026, + "cuda_time_us": 62.232, + "pct_cuda_time": 1.0790425733896158, "trace": "_rocm_C::wvSplitK(bfloat16[28672, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17899,9 +17899,9 @@ { "entry": { "name": "SiluAndMul", - "cpu_time_us": 18.558, - "cuda_time_us": 8.211, - "pct_cuda_time": 0.14306538005977698, + "cpu_time_us": 22.194, + "cuda_time_us": 8.331, + "pct_cuda_time": 0.1444514667519747, "trace": "" }, "children": [ @@ -17909,8 +17909,8 @@ "entry": { "name": "_ZN4vllm18act_and_mul_kernelIN3c108BFloat16ETnPFT_RKS3_EXadL_ZNS_11silu_kernelIS2_EES3_S5_EELb1EEEvPS3_PS4_i", "cpu_time_us": 0, - "cuda_time_us": 8.211, - "pct_cuda_time": 0.14306538005977698, + "cuda_time_us": 8.331, + "pct_cuda_time": 0.1444514667519747, "trace": "_C::silu_and_mul(bfloat16[1, 14336], bfloat16[1, 28672])" }, "children": [] @@ -17920,9 +17920,9 @@ { "entry": { "name": "RowParallelLinear(weight=bfloat16[4096, 14336])", - "cpu_time_us": 47.101, - "cuda_time_us": 25.798, - "pct_cuda_time": 0.4494946626211334, + "cpu_time_us": 52.409, + "cuda_time_us": 28.241, + "pct_cuda_time": 0.4896715727454708, "trace": "" }, "children": [ @@ -17930,8 +17930,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 25.798, - "pct_cuda_time": 0.4494946626211334, + "cuda_time_us": 28.241, + "pct_cuda_time": 0.4896715727454708, "trace": "_rocm_C::wvSplitK(bfloat16[4096, 14336], bfloat16[1, 14336], 304)" }, "children": [] @@ -17945,9 +17945,9 @@ { "entry": { "name": "RMSNorm(weight=bfloat16[4096])", - "cpu_time_us": 24.777, - "cuda_time_us": 4.326, - "pct_cuda_time": 0.07537459921307942, + "cpu_time_us": 28.403, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.07917001526701674, "trace": "" }, "children": [ @@ -17955,8 +17955,8 @@ "entry": { "name": "std::enable_if<(((8)>(0)))&&_typeConvert::exists, void>::type vllm::fused_add_rms_norm_kernel(c10::BFloat16*, c10::BFloat16*, c10::BFloat16 const*, float, int, int)", "cpu_time_us": 0, - "cuda_time_us": 4.326, - "pct_cuda_time": 0.07537459921307942, + "cuda_time_us": 4.566, + "pct_cuda_time": 0.07917001526701674, "trace": "_C::fused_add_rms_norm(bfloat16[1, 4096], bfloat16[1, 4096], bfloat16[4096], 1e-05)" }, "children": [] @@ -17968,9 +17968,9 @@ { "entry": { "name": "LogitsProcessor", - "cpu_time_us": 120.139, - "cuda_time_us": 261.408, - "pct_cuda_time": 4.554674810701033, + "cpu_time_us": 147.2, + "cuda_time_us": 265.255, + "pct_cuda_time": 4.599264651697881, "trace": "" }, "children": [ @@ -17978,8 +17978,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::indexSelectSmallIndex(at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, at::cuda::detail::TensorInfo, int, int, unsigned int, long)", "cpu_time_us": 0, - "cuda_time_us": 3.123, - "pct_cuda_time": 0.05441397904356151, + "cuda_time_us": 3.164, + "pct_cuda_time": 0.05486069389067915, "trace": "index_select(bfloat16[1, 4096], 0, int64[1])" }, "children": [] @@ -17988,8 +17988,8 @@ "entry": { "name": "void wvSplitK_hf_sml_<__hip_bfloat16, 64, 2, 16, 8, 2, 1>(int, int, __hip_bfloat16 const*, __hip_bfloat16 const*, __hip_bfloat16*, int, int)", "cpu_time_us": 0, - "cuda_time_us": 258.285, - "pct_cuda_time": 4.500260831657472, + "cuda_time_us": 262.091, + "pct_cuda_time": 4.5444039578072015, "trace": "_rocm_C::wvSplitK(bfloat16[128256, 4096], bfloat16[1, 4096], 304)" }, "children": [] @@ -17999,9 +17999,9 @@ { "entry": { "name": "Sampler", - "cpu_time_us": 875.007, - "cuda_time_us": 185.66100000000003, - "pct_cuda_time": 3.2348875322467734, + "cpu_time_us": 963.44, + "cuda_time_us": 185.664, + "pct_cuda_time": 3.2192338402399026, "trace": "" }, "children": [ @@ -18009,8 +18009,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 4.886, - "pct_cuda_time": 0.08513182888467548, + "cuda_time_us": 2.483, + "pct_cuda_time": 0.043052813821288345, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -18019,8 +18019,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 2.442, - "pct_cuda_time": 0.042548490817924174, + "cuda_time_us": 2.322, + "pct_cuda_time": 0.040261229840125465, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -18029,8 +18029,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 2.483, - "pct_cuda_time": 0.04326285941888031, + "cuda_time_us": 2.724, + "pct_cuda_time": 0.047231520277563215, "trace": "copy_(int32[1], int32[1], True) <- _to_copy(int32[1], 3, 0, None, None, True, None) <- to(int32[1], 3, 0, None, None, True, False, None)" }, "children": [] @@ -18039,8 +18039,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 2.362, - "pct_cuda_time": 0.041154600864839026, + "cuda_time_us": 2.483, + "pct_cuda_time": 0.043052813821288345, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -18049,8 +18049,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 2.323, - "pct_cuda_time": 0.04047507951271001, + "cuda_time_us": 2.363, + "pct_cuda_time": 0.04097213010862036, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -18059,8 +18059,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 2.282, - "pct_cuda_time": 0.03976071091175388, + "cuda_time_us": 2.363, + "pct_cuda_time": 0.04097213010862036, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -18069,8 +18069,8 @@ "entry": { "name": "Memcpy DtoD (Device -> Device)", "cpu_time_us": 0, - "cuda_time_us": 2.322, - "pct_cuda_time": 0.04045765588829645, + "cuda_time_us": 2.402, + "pct_cuda_time": 0.041648352315237454, "trace": "copy_(bfloat16[1], bfloat16[1], True) <- _to_copy(bfloat16[1], 15, 0, None, None, True, None) <- to(bfloat16[1], 15, 0, None, None, True, False, None)" }, "children": [] @@ -18079,8 +18079,8 @@ "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#7}::operator()() const::{lambda(float)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 3.164, - "pct_cuda_time": 0.05512834764451765, + "cuda_time_us": 3.524, + "pct_cuda_time": 0.06110274502868309, "trace": "copy_(float32[1, 128256], bfloat16[1, 128256], False) <- _to_copy(bfloat16[1, 128256], 6, None, None, None, False, None) <- to(bfloat16[1, 128256], 6, False, False, None)" }, "children": [] @@ -18089,8 +18089,8 @@ "entry": { "name": "void at::native::elementwise_kernel<128, 4, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2}>(int, at::native::gpu_kernel_impl > >(at::TensorIteratorBase&, at::native::BinaryFunctor > const&)::{lambda(int)#2})", "cpu_time_us": 0, - "cuda_time_us": 9.333, - "pct_cuda_time": 0.1626146866517962, + "cuda_time_us": 10.254, + "pct_cuda_time": 0.17779442324747913, "trace": "div_(float32[1, 128256], bfloat16[1, 1])" }, "children": [] @@ -18099,8 +18099,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::SoftMaxForwardEpilogue>(float*, float const*, int)", "cpu_time_us": 0, - "cuda_time_us": 63.494, - "pct_cuda_time": 1.1062956085148556, + "cuda_time_us": 64.015, + "pct_cuda_time": 1.1099580655536743, "trace": "_softmax(float32[1, 128256], -1, False) <- softmax(float32[1, 128256], -1, 6)" }, "children": [] @@ -18109,8 +18109,8 @@ "entry": { "name": "void at::native::(anonymous namespace)::cunn_SoftMaxForward<4, float, float, float, at::native::(anonymous namespace)::LogSoftMaxForwardEpilogue>(float*, float const*, int)", "cpu_time_us": 0, - "cuda_time_us": 35.973, - "pct_cuda_time": 0.6267800410291507, + "cuda_time_us": 35.733, + "pct_cuda_time": 0.6195755925397085, "trace": "_log_softmax(float32[1, 128256], -1, False) <- log_softmax(float32[1, 128256], -1, 6)" }, "children": [] @@ -18119,8 +18119,8 @@ "entry": { "name": "void at::native::elementwise_kernel<512, 1, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1}>(int, at::native::gpu_kernel_impl(at::TensorIteratorBase&, at::native::direct_copy_kernel_cuda(at::TensorIteratorBase&)::{lambda()#3}::operator()() const::{lambda()#4}::operator()() const::{lambda(long)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 3.804, - "pct_cuda_time": 0.06627946726919884, + "cuda_time_us": 3.844, + "pct_cuda_time": 0.06665123492913105, "trace": "copy_(int64[1], int32[1], False) <- _to_copy(int32[1], 4, None, None, None, False, None) <- to(int32[1], 4, False, False, None)" }, "children": [] @@ -18129,8 +18129,8 @@ "entry": { "name": "void at::native::index_elementwise_kernel<128, 4, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1}>(long, at::native::gpu_index_kernel >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1}>(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef, at::native::index_kernel_impl >(at::TensorIteratorBase&, c10::ArrayRef, c10::ArrayRef)::{lambda(char*, char const*, long)#1} const&)::{lambda(int)#1})", "cpu_time_us": 0, - "cuda_time_us": 5.567, - "pct_cuda_time": 0.0969973171103128, + "cuda_time_us": 5.848, + "pct_cuda_time": 0.10139865293068634, "trace": "index(float32[1, 128256], None)" }, "children": [] @@ -18139,8 +18139,8 @@ "entry": { "name": "void at::native::reduce_kernel<512, 1, at::native::ReduceOp, unsigned int, long, 4, 4> >(at::native::ReduceOp, unsigned int, long, 4, 4>)", "cpu_time_us": 0, - "cuda_time_us": 40.78, - "pct_cuda_time": 0.7105354035851547, + "cuda_time_us": 40.58, + "pct_cuda_time": 0.7036178755005561, "trace": "argmax(float32[1, 128256], -1, False)" }, "children": [] @@ -18149,8 +18149,8 @@ "entry": { "name": "Memcpy DtoH (Device -> Host)", "cpu_time_us": 0, - "cuda_time_us": 4.446, - "pct_cuda_time": 0.07746543414270715, + "cuda_time_us": 4.726, + "pct_cuda_time": 0.08194426021724072, "trace": "copy_(int64[1], int64[1], False) <- _to_copy(int64[1], 4, 0, None, None, False, None) <- to(int64[1], 4, 0, None, None, False, False, None)" }, "children": []