Lines Matching refs:xgmi
324 return sysfs_emit(buf, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
485 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
496 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
548 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
619 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
620 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
651 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
652 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
672 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
684 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_num_links()
701 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_initialize_hive_get_data_partition()
724 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device()
727 if (!adev->gmc.xgmi.pending_reset && in amdgpu_xgmi_add_device()
736 ret = psp_xgmi_get_hive_id(&adev->psp, &adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
743 ret = psp_xgmi_get_node_id(&adev->psp, &adev->gmc.xgmi.node_id); in amdgpu_xgmi_add_device()
750 adev->gmc.xgmi.hive_id = 16; in amdgpu_xgmi_add_device()
751 adev->gmc.xgmi.node_id = adev->gmc.xgmi.physical_node_id + 16; in amdgpu_xgmi_add_device()
759 adev->gmc.xgmi.node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
766 list_add_tail(&adev->gmc.xgmi.head, &hive->device_list); in amdgpu_xgmi_add_device()
774 if (!adev->gmc.xgmi.pending_reset && in amdgpu_xgmi_add_device()
776 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
781 adev->gmc.xgmi.node_id; in amdgpu_xgmi_add_device()
790 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
796 tmp_adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
797 tmp_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
812 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
818 tmp_adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
819 tmp_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
832 if (!ret && !adev->gmc.xgmi.pending_reset) in amdgpu_xgmi_add_device()
841 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
845 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id, in amdgpu_xgmi_add_device()
856 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_remove_device()
867 list_del(&adev->gmc.xgmi.head); in amdgpu_xgmi_remove_device()
887 if (!adev->gmc.xgmi.supported || in amdgpu_xgmi_ras_late_init()
888 adev->gmc.xgmi.num_physical_nodes == 0) in amdgpu_xgmi_ras_late_init()
891 adev->gmc.xgmi.ras->ras_block.hw_ops->reset_ras_error_count(adev); in amdgpu_xgmi_ras_late_init()
899 struct amdgpu_xgmi *xgmi = &adev->gmc.xgmi; in amdgpu_xgmi_get_relative_phy_addr() local
900 return (addr + xgmi->physical_node_id * xgmi->node_segment_size); in amdgpu_xgmi_get_relative_phy_addr()
1057 adev->gmc.xgmi.ras->ras_block.hw_ops->reset_ras_error_count(adev); in amdgpu_xgmi_query_ras_error_count()
1109 if (!adev->gmc.xgmi.ras) in amdgpu_xgmi_ras_sw_init()
1112 ras = adev->gmc.xgmi.ras; in amdgpu_xgmi_ras_sw_init()
1122 adev->gmc.xgmi.ras_if = &ras->ras_block.ras_comm; in amdgpu_xgmi_ras_sw_init()