Skip to content

Instantly share code, notes, and snippets.

@tpruzina
Last active February 25, 2017 12:47
Show Gist options
  • Save tpruzina/4a4aef53b266aa04606385b92ca6d042 to your computer and use it in GitHub Desktop.
Save tpruzina/4a4aef53b266aa04606385b92ca6d042 to your computer and use it in GitHub Desktop.
NVIDIA-378.13-KERNEL-4.10-stageing-7bb033829ef3ecfc491c0ed0197966e8f197fbdc.patch
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/common/inc/nv-mm.h NVIDIA-Linux-x86_64-378.13.patched/kernel/common/inc/nv-mm.h
--- NVIDIA-Linux-x86_64-378.13/kernel/common/inc/nv-mm.h 2017-02-08 04:58:34.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/common/inc/nv-mm.h 2017-02-24 21:38:54.126260185 +0100
@@ -46,6 +46,8 @@
* 2016 Dec 14:5b56d49fc31dbb0487e14ead790fc81ca9fb2c99
*/
+#include <linux/version.h>
+
#if defined(NV_GET_USER_PAGES_REMOTE_PRESENT)
#if defined(NV_GET_USER_PAGES_HAS_WRITE_AND_FORCE_ARGS)
#define NV_GET_USER_PAGES get_user_pages
@@ -92,10 +94,13 @@
pages, vmas, NULL);
#else
-
+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 10, 0)
return get_user_pages_remote(tsk, mm, start, nr_pages, flags,
pages, vmas);
-
+#else
+ return get_user_pages_remote(tsk, mm, start, nr_pages, flags,
+ pages, vmas, NULL);
+#endif
#endif
}
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia/nv-p2p.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia/nv-p2p.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia/nv-p2p.c 2017-02-08 04:58:34.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia/nv-p2p.c 2017-02-24 21:38:54.126260185 +0100
@@ -146,7 +146,7 @@
int nvidia_p2p_get_pages(
uint64_t p2p_token,
uint32_t va_space,
- uint64_t virtual_address,
+ uint64_t address,
uint64_t length,
struct nvidia_p2p_page_table **page_table,
void (*free_callback)(void * data),
@@ -211,7 +211,7 @@
}
status = rm_p2p_get_pages(sp, p2p_token, va_space,
- virtual_address, length, physical_addresses, wreqmb_h,
+ address, length, physical_addresses, wreqmb_h,
rreqmb_h, &entries, &gpu_uuid, *page_table,
free_callback, data);
if (status != NV_OK)
@@ -286,7 +286,7 @@
if (bGetPages)
{
- rm_p2p_put_pages(sp, p2p_token, va_space, virtual_address,
+ rm_p2p_put_pages(sp, p2p_token, va_space, address,
gpu_uuid, *page_table);
}
@@ -329,7 +329,7 @@
int nvidia_p2p_put_pages(
uint64_t p2p_token,
uint32_t va_space,
- uint64_t virtual_address,
+ uint64_t address,
struct nvidia_p2p_page_table *page_table
)
{
@@ -343,7 +343,7 @@
return rc;
}
- status = rm_p2p_put_pages(sp, p2p_token, va_space, virtual_address,
+ status = rm_p2p_put_pages(sp, p2p_token, va_space, address,
page_table->gpu_uuid, page_table);
if (status == NV_OK)
nvidia_p2p_free_page_table(page_table);
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia/nv-pat.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia/nv-pat.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia/nv-pat.c 2017-02-08 04:58:34.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia/nv-pat.c 2017-02-24 21:38:54.126260185 +0100
@@ -203,6 +203,7 @@
}
#if defined(NV_ENABLE_PAT_SUPPORT) && defined(NV_ENABLE_HOTPLUG_CPU)
+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 10, 0)
static int
nvidia_cpu_callback(struct notifier_block *nfb, unsigned long action, void *hcpu)
{
@@ -234,6 +235,34 @@
.notifier_call = nvidia_cpu_callback,
.priority = 0
};
+#else
+static int nvidia_cpu_online(unsigned int hcpu)
+{
+ unsigned int cpu = get_cpu();
+ if (cpu == hcpu)
+ nv_setup_pat_entries(NULL);
+ else
+ NV_SMP_CALL_FUNCTION(nv_setup_pat_entries, (void *)(long int)hcpu, 1);
+
+ put_cpu();
+
+ return 0;
+}
+
+static int nvidia_cpu_down_prep(unsigned int hcpu)
+{
+ unsigned int cpu = get_cpu();
+ if (cpu == hcpu)
+ nv_restore_pat_entries(NULL);
+ else
+ NV_SMP_CALL_FUNCTION(nv_restore_pat_entries, (void *)(long int)hcpu, 1);
+
+ put_cpu();
+
+ return 0;
+}
+#endif
+
#endif
int nv_init_pat_support(nvidia_stack_t *sp)
@@ -255,7 +284,14 @@
#if defined(NV_ENABLE_PAT_SUPPORT) && defined(NV_ENABLE_HOTPLUG_CPU)
if (nv_pat_mode == NV_PAT_MODE_BUILTIN)
{
+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 10, 0)
if (register_hotcpu_notifier(&nv_hotcpu_nfb) != 0)
+#else
+ if (cpuhp_setup_state(CPUHP_AP_ONLINE_DYN,
+ "gpu/nvidia:online",
+ nvidia_cpu_online,
+ nvidia_cpu_down_prep) != 0)
+#endif
{
nv_disable_pat_support();
nv_printf(NV_DBG_ERRORS,
@@ -280,7 +316,11 @@
{
nv_disable_pat_support();
#if defined(NV_ENABLE_PAT_SUPPORT) && defined(NV_ENABLE_HOTPLUG_CPU)
+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 10, 0)
unregister_hotcpu_notifier(&nv_hotcpu_nfb);
+#else
+ cpuhp_remove_state_nocalls(CPUHP_AP_ONLINE_DYN);
+#endif
#endif
}
}
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-drv.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-drv.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-drv.c 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-drv.c 2017-02-25 13:40:38.362643797 +0100
@@ -71,12 +71,8 @@
static struct drm_framebuffer *nvidia_drm_framebuffer_create(
struct drm_device *dev,
struct drm_file *file,
- #if defined(NV_DRM_HELPER_MODE_FILL_FB_STRUCT_HAS_CONST_MODE_CMD_ARG)
const struct drm_mode_fb_cmd2 *cmd
- #else
- struct drm_mode_fb_cmd2 *cmd
- #endif
- )
+)
{
struct drm_mode_fb_cmd2 local_cmd;
struct drm_framebuffer *fb;
@@ -88,14 +84,10 @@
file,
&local_cmd, 0, 0);
- #if !defined(NV_DRM_HELPER_MODE_FILL_FB_STRUCT_HAS_CONST_MODE_CMD_ARG)
- *cmd = local_cmd;
- #endif
-
return fb;
}
-static const struct drm_mode_config_funcs nv_mode_config_funcs = {
+struct drm_mode_config_funcs nv_mode_config_funcs = {
.fb_create = nvidia_drm_framebuffer_create,
#if defined(NV_DRM_MODE_CONFIG_FUNCS_HAS_ATOMIC_STATE_ALLOC)
@@ -398,7 +390,8 @@
return 0;
}
-static int nvidia_drm_unload(struct drm_device *dev)
+
+static void nvidia_drm_unload(struct drm_device *dev)
{
#if defined(NV_DRM_ATOMIC_MODESET_AVAILABLE)
struct NvKmsKapiDevice *pDevice = NULL;
@@ -412,7 +405,7 @@
if (!nvidia_drm_modeset_enabled(dev))
{
- return 0;
+ return;
}
mutex_lock(&nv_dev->lock);
@@ -445,7 +438,7 @@
#endif /* NV_DRM_ATOMIC_MODESET_AVAILABLE */
- return 0;
+ return;
}
#if defined(NV_DRM_ATOMIC_MODESET_AVAILABLE)
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-encoder.h NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-encoder.h
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-encoder.h 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-encoder.h 2017-02-24 21:45:22.497238703 +0100
@@ -30,6 +30,7 @@
#include "nvidia-drm-priv.h"
#include <drm/drmP.h>
+#include <drm/drm_encoder.h>
#include "nvkms-kapi.h"
struct nvidia_drm_encoder
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-fb.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-fb.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-fb.c 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-fb.c 2017-02-24 22:19:05.963126779 +0100
@@ -152,7 +152,7 @@
/* Fill out framebuffer metadata from the userspace fb creation request */
- drm_helper_mode_fill_fb_struct(&nv_fb->base, cmd);
+ drm_helper_mode_fill_fb_struct(dev, &nv_fb->base, cmd);
/* Initialize the base framebuffer object and add it to drm subsystem */
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-fence.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-fence.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-fence.c 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-fence.c 2017-02-24 21:38:54.126260185 +0100
@@ -31,7 +31,7 @@
#if defined(NV_DRM_DRIVER_HAS_GEM_PRIME_RES_OBJ)
struct nv_fence {
- struct fence base;
+ struct dma_fence base;
spinlock_t lock;
struct nvidia_drm_device *nv_dev;
@@ -51,7 +51,7 @@
static const char *nvidia_drm_gem_prime_fence_op_get_driver_name
(
- struct fence *fence
+ struct dma_fence *fence
)
{
return "NVIDIA";
@@ -59,7 +59,7 @@
static const char *nvidia_drm_gem_prime_fence_op_get_timeline_name
(
- struct fence *fence
+ struct dma_fence *fence
)
{
return "nvidia.prime";
@@ -67,7 +67,7 @@
static bool nvidia_drm_gem_prime_fence_op_signaled
(
- struct fence *fence
+ struct dma_fence *fence
)
{
struct nv_fence *nv_fence = container_of(fence, struct nv_fence, base);
@@ -99,7 +99,7 @@
static bool nvidia_drm_gem_prime_fence_op_enable_signaling
(
- struct fence *fence
+ struct dma_fence *fence
)
{
bool ret = true;
@@ -107,7 +107,7 @@
struct nvidia_drm_gem_object *nv_gem = nv_fence->nv_gem;
struct nvidia_drm_device *nv_dev = nv_fence->nv_dev;
- if (fence_is_signaled(fence))
+ if (dma_fence_is_signaled(fence))
{
return false;
}
@@ -136,7 +136,7 @@
}
nv_gem->fenceContext.softFence = fence;
- fence_get(fence);
+ dma_fence_get(fence);
unlock_struct_mutex:
mutex_unlock(&nv_dev->dev->struct_mutex);
@@ -146,7 +146,7 @@
static void nvidia_drm_gem_prime_fence_op_release
(
- struct fence *fence
+ struct dma_fence *fence
)
{
struct nv_fence *nv_fence = container_of(fence, struct nv_fence, base);
@@ -155,7 +155,7 @@
static signed long nvidia_drm_gem_prime_fence_op_wait
(
- struct fence *fence,
+ struct dma_fence *fence,
bool intr,
signed long timeout
)
@@ -170,12 +170,12 @@
* that it should never get hit during normal operation, but not so long
* that the system becomes unresponsive.
*/
- return fence_default_wait(fence, intr,
+ return dma_fence_default_wait(fence, intr,
(timeout == MAX_SCHEDULE_TIMEOUT) ?
msecs_to_jiffies(96) : timeout);
}
-static const struct fence_ops nvidia_drm_gem_prime_fence_ops = {
+static const struct dma_fence_ops nvidia_drm_gem_prime_fence_ops = {
.get_driver_name = nvidia_drm_gem_prime_fence_op_get_driver_name,
.get_timeline_name = nvidia_drm_gem_prime_fence_op_get_timeline_name,
.signaled = nvidia_drm_gem_prime_fence_op_signaled,
@@ -285,7 +285,7 @@
bool force
)
{
- struct fence *fence = nv_gem->fenceContext.softFence;
+ struct dma_fence *fence = nv_gem->fenceContext.softFence;
WARN_ON(!mutex_is_locked(&nv_dev->dev->struct_mutex));
@@ -301,10 +301,10 @@
if (force || nv_fence_ready_to_signal(nv_fence))
{
- fence_signal(&nv_fence->base);
+ dma_fence_signal(&nv_fence->base);
nv_gem->fenceContext.softFence = NULL;
- fence_put(&nv_fence->base);
+ dma_fence_put(&nv_fence->base);
nvKms->disableChannelEvent(nv_dev->pDevice,
nv_gem->fenceContext.cb);
@@ -320,7 +320,7 @@
nv_fence = container_of(fence, struct nv_fence, base);
- fence_signal(&nv_fence->base);
+ dma_fence_signal(&nv_fence->base);
}
}
@@ -513,7 +513,7 @@
* fence_context_alloc() cannot fail, so we do not need to check a return
* value.
*/
- nv_gem->fenceContext.context = fence_context_alloc(1);
+ nv_gem->fenceContext.context = dma_fence_context_alloc(1);
ret = nvidia_drm_gem_prime_fence_import_semaphore(
nv_dev, nv_gem, p->index,
@@ -670,7 +670,7 @@
nv_fence->nv_gem = nv_gem;
spin_lock_init(&nv_fence->lock);
- fence_init(&nv_fence->base, &nvidia_drm_gem_prime_fence_ops,
+ dma_fence_init(&nv_fence->base, &nvidia_drm_gem_prime_fence_ops,
&nv_fence->lock, nv_gem->fenceContext.context,
p->sem_thresh);
@@ -680,7 +680,7 @@
reservation_object_add_excl_fence(&nv_gem->fenceContext.resv,
&nv_fence->base);
- fence_put(&nv_fence->base); /* Reservation object has reference */
+ dma_fence_put(&nv_fence->base); /* Reservation object has reference */
ret = 0;
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-gem.h NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-gem.h
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-gem.h 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-gem.h 2017-02-24 21:38:54.127260185 +0100
@@ -29,6 +29,8 @@
#include "nvidia-drm-priv.h"
+#include <linux/version.h>
+
#include <drm/drmP.h>
#include "nvkms-kapi.h"
@@ -101,7 +103,7 @@
/* Software signaling structures */
struct NvKmsKapiChannelEvent *cb;
struct nvidia_drm_gem_prime_soft_fence_event_args *cbArgs;
- struct fence *softFence; /* Fence for software signaling */
+ struct dma_fence *softFence; /* Fence for software signaling */
} fenceContext;
#endif
};
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-modeset.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-modeset.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-modeset.c 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-modeset.c 2017-02-24 21:38:54.127260185 +0100
@@ -78,8 +78,7 @@
void nvidia_drm_atomic_state_free(struct drm_atomic_state *state)
{
- struct nvidia_drm_atomic_state *nv_state =
- to_nv_atomic_state(state);
+ struct nvidia_drm_atomic_state *nv_state = to_nv_atomic_state(state);
drm_atomic_state_default_release(state);
nvidia_drm_free(nv_state);
}
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-priv.h NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-priv.h
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-drm/nvidia-drm-priv.h 2017-02-08 04:58:37.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-drm/nvidia-drm-priv.h 2017-02-24 21:38:54.128260185 +0100
@@ -25,6 +25,8 @@
#include "conftest.h" /* NV_DRM_AVAILABLE */
+#include <linux/version.h>
+
#if defined(NV_DRM_AVAILABLE)
#include <drm/drmP.h>
@@ -34,7 +36,7 @@
#endif
#if defined(NV_DRM_DRIVER_HAS_GEM_PRIME_RES_OBJ)
-#include <linux/fence.h>
+#include <linux/dma-fence.h>
#include <linux/reservation.h>
#endif
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-uvm/uvm8_gpu_isr.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-uvm/uvm8_gpu_isr.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-uvm/uvm8_gpu_isr.c 2017-02-08 04:58:35.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-uvm/uvm8_gpu_isr.c 2017-02-24 21:38:54.128260185 +0100
@@ -270,7 +270,7 @@
void uvm_gpu_replayable_faults_isr_unlock(uvm_gpu_t *gpu)
{
- UVM_ASSERT(atomic_read(&gpu->gpu_kref.refcount) > 0);
+ UVM_ASSERT(atomic_read(&(gpu->gpu_kref.refcount.refs)) > 0);
uvm_spin_lock_irqsave(&gpu->isr.replayable_faults.interrupts_lock);
diff -ur NVIDIA-Linux-x86_64-378.13/kernel/nvidia-uvm/uvm8_test.c NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-uvm/uvm8_test.c
--- NVIDIA-Linux-x86_64-378.13/kernel/nvidia-uvm/uvm8_test.c 2017-02-08 04:58:36.000000000 +0100
+++ NVIDIA-Linux-x86_64-378.13.patched/kernel/nvidia-uvm/uvm8_test.c 2017-02-24 21:38:54.129260185 +0100
@@ -103,7 +103,7 @@
return NV_ERR_INVALID_STATE;
}
-static NV_STATUS uvm8_test_get_kernel_virtual_address(
+static NV_STATUS uvm8_test_get_kernel_address(
UVM_TEST_GET_KERNEL_VIRTUAL_ADDRESS_PARAMS *params,
struct file *filp)
{
@@ -173,7 +173,7 @@
UVM_ROUTE_CMD_STACK(UVM_TEST_RANGE_GROUP_RANGE_COUNT, uvm8_test_range_group_range_count);
UVM_ROUTE_CMD_STACK(UVM_TEST_GET_PREFETCH_FAULTS_REENABLE_LAPSE, uvm8_test_get_prefetch_faults_reenable_lapse);
UVM_ROUTE_CMD_STACK(UVM_TEST_SET_PREFETCH_FAULTS_REENABLE_LAPSE, uvm8_test_set_prefetch_faults_reenable_lapse);
- UVM_ROUTE_CMD_STACK(UVM_TEST_GET_KERNEL_VIRTUAL_ADDRESS, uvm8_test_get_kernel_virtual_address);
+ UVM_ROUTE_CMD_STACK(UVM_TEST_GET_KERNEL_VIRTUAL_ADDRESS, uvm8_test_get_kernel_address);
UVM_ROUTE_CMD_STACK(UVM_TEST_PMA_ALLOC_FREE, uvm8_test_pma_alloc_free);
UVM_ROUTE_CMD_STACK(UVM_TEST_PMM_ALLOC_FREE_ROOT, uvm8_test_pmm_alloc_free_root);
UVM_ROUTE_CMD_STACK(UVM_TEST_PMM_INJECT_PMA_EVICT_ERROR, uvm8_test_pmm_inject_pma_evict_error);
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment