1 | /* Most of this code is ripped from the Linux kernel:
|
---|
2 | *
|
---|
3 | * linux/drivers/char/mem.c
|
---|
4 | *
|
---|
5 | * Copyright (C) 1991, 1992 Linus Torvalds
|
---|
6 | *
|
---|
7 | * Added devfs support.
|
---|
8 | * Jan-11-1998, C. Scott Ananian <cananian@alumni.princeton.edu>
|
---|
9 | * Shared /dev/zero mmaping support, Feb 2000, Kanoj Sarcar <kanoj@sgi.com>
|
---|
10 | */
|
---|
11 |
|
---|
12 | #include "config.h"
|
---|
13 |
|
---|
14 | #undef _
|
---|
15 | #define _(string) string
|
---|
16 |
|
---|
17 | #include <linux/module.h>
|
---|
18 | #include <linux/kernel.h>
|
---|
19 | #include <linux/proc_fs.h>
|
---|
20 | #include <linux/string.h>
|
---|
21 | #include <linux/vmalloc.h>
|
---|
22 | #include <linux/mm.h>
|
---|
23 | #include <linux/miscdevice.h>
|
---|
24 | #include <linux/slab.h>
|
---|
25 | #include <linux/vmalloc.h>
|
---|
26 | #include <linux/mman.h>
|
---|
27 | #include <linux/random.h>
|
---|
28 | #include <linux/init.h>
|
---|
29 | #include <linux/raw.h>
|
---|
30 | #include <linux/tty.h>
|
---|
31 | #include <linux/capability.h>
|
---|
32 | #include <linux/ptrace.h>
|
---|
33 | #include <linux/device.h>
|
---|
34 | #include <linux/highmem.h>
|
---|
35 | #include <linux/crash_dump.h>
|
---|
36 | #include <linux/backing-dev.h>
|
---|
37 | #include <linux/bootmem.h>
|
---|
38 | #include <linux/splice.h>
|
---|
39 | #include <linux/pfn.h>
|
---|
40 | #include <linux/version.h>
|
---|
41 |
|
---|
42 | #include <asm/uaccess.h>
|
---|
43 | #include <asm/io.h>
|
---|
44 | #include <asm/pgtable.h>
|
---|
45 |
|
---|
46 | #ifdef CONFIG_IA64
|
---|
47 | # include <linux/efi.h>
|
---|
48 | #endif
|
---|
49 |
|
---|
50 | MODULE_LICENSE("GPL");
|
---|
51 | MODULE_DESCRIPTION("samhain_kmem Kernel Module");
|
---|
52 | MODULE_AUTHOR("Rainer Wichmann");
|
---|
53 |
|
---|
54 | static int debug = 0;
|
---|
55 | #ifdef MODULE_PARM
|
---|
56 | MODULE_PARM (debug, "i");
|
---|
57 | #else
|
---|
58 | module_param(debug, int, 0444);
|
---|
59 | #endif
|
---|
60 |
|
---|
61 | #ifdef MODULE_PARM_DESC
|
---|
62 | MODULE_PARM_DESC(debug, "Set to a non-zero value for debugging.");
|
---|
63 | #endif
|
---|
64 |
|
---|
65 | /* struct task_struct
|
---|
66 | */
|
---|
67 | #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,29)
|
---|
68 | #define TASK_EUID euid
|
---|
69 | #else
|
---|
70 | #define TASK_EUID cred->euid
|
---|
71 | #endif
|
---|
72 |
|
---|
73 | static struct proc_dir_entry *proc_entry;
|
---|
74 |
|
---|
75 | /*
|
---|
76 | * Architectures vary in how they handle caching for addresses
|
---|
77 | * outside of main memory.
|
---|
78 | *
|
---|
79 | */
|
---|
80 | static inline int uncached_access(struct file *file, unsigned long addr)
|
---|
81 | {
|
---|
82 | #if defined(__i386__) && !defined(__arch_um__)
|
---|
83 | /*
|
---|
84 | * On the PPro and successors, the MTRRs are used to set
|
---|
85 | * memory types for physical addresses outside main memory,
|
---|
86 | * so blindly setting PCD or PWT on those pages is wrong.
|
---|
87 | * For Pentiums and earlier, the surround logic should disable
|
---|
88 | * caching for the high addresses through the KEN pin, but
|
---|
89 | * we maintain the tradition of paranoia in this code.
|
---|
90 | */
|
---|
91 | if (file->f_flags & O_SYNC)
|
---|
92 | return 1;
|
---|
93 | return !( test_bit(X86_FEATURE_MTRR, (const void *) boot_cpu_data.x86_capability) ||
|
---|
94 | test_bit(X86_FEATURE_K6_MTRR, (const void *) boot_cpu_data.x86_capability) ||
|
---|
95 | test_bit(X86_FEATURE_CYRIX_ARR, (const void *) boot_cpu_data.x86_capability) ||
|
---|
96 | test_bit(X86_FEATURE_CENTAUR_MCR, (const void *) boot_cpu_data.x86_capability) )
|
---|
97 | && addr >= __pa(high_memory);
|
---|
98 | #elif defined(__x86_64__) && !defined(__arch_um__)
|
---|
99 | /*
|
---|
100 | * This is broken because it can generate memory type aliases,
|
---|
101 | * which can cause cache corruptions
|
---|
102 | * But it is only available for root and we have to be bug-to-bug
|
---|
103 | * compatible with i386.
|
---|
104 | */
|
---|
105 | if (file->f_flags & O_SYNC)
|
---|
106 | return 1;
|
---|
107 | /* same behaviour as i386. PAT always set to cached and MTRRs control the
|
---|
108 | caching behaviour.
|
---|
109 | Hopefully a full PAT implementation will fix that soon. */
|
---|
110 | return 0;
|
---|
111 | #elif defined(CONFIG_IA64)
|
---|
112 | /*
|
---|
113 | * On ia64, we ignore O_SYNC because we cannot tolerate
|
---|
114 | * memory attribute aliases.
|
---|
115 | */
|
---|
116 | return !(efi_mem_attributes(addr) & EFI_MEMORY_WB);
|
---|
117 | #elif defined(CONFIG_MIPS)
|
---|
118 | {
|
---|
119 | extern int __uncached_access(struct file *file,
|
---|
120 | unsigned long addr);
|
---|
121 |
|
---|
122 | return __uncached_access(file, addr);
|
---|
123 | }
|
---|
124 | #else
|
---|
125 | /*
|
---|
126 | * Accessing memory above the top the kernel knows about
|
---|
127 | * or through a file pointer
|
---|
128 | * that was marked O_SYNC will be done non-cached.
|
---|
129 | */
|
---|
130 | if (file->f_flags & O_SYNC)
|
---|
131 | return 1;
|
---|
132 | return addr >= __pa(high_memory);
|
---|
133 | #endif
|
---|
134 | }
|
---|
135 |
|
---|
136 | #ifndef ARCH_HAS_VALID_PHYS_ADDR_RANGE
|
---|
137 | static inline int valid_phys_addr_range(unsigned long addr, size_t count)
|
---|
138 | {
|
---|
139 | if (addr + count > __pa(high_memory))
|
---|
140 | return 0;
|
---|
141 |
|
---|
142 | return 1;
|
---|
143 | }
|
---|
144 |
|
---|
145 | static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size)
|
---|
146 | {
|
---|
147 | return 1;
|
---|
148 | }
|
---|
149 | #endif
|
---|
150 |
|
---|
151 |
|
---|
152 | /* #ifndef __HAVE_PHYS_MEM_ACCESS_PROT */
|
---|
153 | static pgprot_t my_phys_mem_access_prot(struct file *file, unsigned long pfn,
|
---|
154 | unsigned long size, pgprot_t vma_prot)
|
---|
155 | {
|
---|
156 | #ifdef pgprot_noncached
|
---|
157 | unsigned long offset = pfn << PAGE_SHIFT;
|
---|
158 |
|
---|
159 | if (uncached_access(file, offset))
|
---|
160 | return pgprot_noncached(vma_prot);
|
---|
161 | #else
|
---|
162 | #error pgtable
|
---|
163 | #endif
|
---|
164 | return vma_prot;
|
---|
165 | }
|
---|
166 | /* #endif */
|
---|
167 |
|
---|
168 |
|
---|
169 | #ifndef CONFIG_MMU
|
---|
170 | static unsigned long get_unmapped_area_mem(struct file *file,
|
---|
171 | unsigned long addr,
|
---|
172 | unsigned long len,
|
---|
173 | unsigned long pgoff,
|
---|
174 | unsigned long flags)
|
---|
175 | {
|
---|
176 | if (!valid_mmap_phys_addr_range(pgoff, len))
|
---|
177 | return (unsigned long) -EINVAL;
|
---|
178 | return pgoff << PAGE_SHIFT;
|
---|
179 | }
|
---|
180 |
|
---|
181 | /* can't do an in-place private mapping if there's no MMU */
|
---|
182 | static inline int private_mapping_ok(struct vm_area_struct *vma)
|
---|
183 | {
|
---|
184 | return vma->vm_flags & VM_MAYSHARE;
|
---|
185 | }
|
---|
186 | #else
|
---|
187 | #define get_unmapped_area_mem NULL
|
---|
188 |
|
---|
189 | static inline int private_mapping_ok(struct vm_area_struct *vma)
|
---|
190 | {
|
---|
191 | return 1;
|
---|
192 | }
|
---|
193 | #endif
|
---|
194 |
|
---|
195 | static int mmap_mem(struct file * file, struct vm_area_struct * vma)
|
---|
196 | {
|
---|
197 | size_t size = vma->vm_end - vma->vm_start;
|
---|
198 |
|
---|
199 | if (!valid_mmap_phys_addr_range(vma->vm_pgoff, size))
|
---|
200 | return -EINVAL;
|
---|
201 |
|
---|
202 | if (!private_mapping_ok(vma))
|
---|
203 | return -ENOSYS;
|
---|
204 |
|
---|
205 | vma->vm_page_prot = my_phys_mem_access_prot(file, vma->vm_pgoff,
|
---|
206 | size,
|
---|
207 | vma->vm_page_prot);
|
---|
208 |
|
---|
209 | /* Remap-pfn-range will mark the range VM_IO and VM_RESERVED */
|
---|
210 | if (remap_pfn_range(vma,
|
---|
211 | vma->vm_start,
|
---|
212 | vma->vm_pgoff,
|
---|
213 | size,
|
---|
214 | vma->vm_page_prot))
|
---|
215 | return -EAGAIN;
|
---|
216 | return 0;
|
---|
217 | }
|
---|
218 |
|
---|
219 | static int mmap_kmem(struct file * file, struct vm_area_struct * vma)
|
---|
220 | {
|
---|
221 | unsigned long pfn;
|
---|
222 |
|
---|
223 | /* Turn a kernel-virtual address into a physical page frame */
|
---|
224 | pfn = __pa((u64)vma->vm_pgoff << PAGE_SHIFT) >> PAGE_SHIFT;
|
---|
225 |
|
---|
226 | /*
|
---|
227 | * RED-PEN: on some architectures there is more mapped memory
|
---|
228 | * than available in mem_map which pfn_valid checks
|
---|
229 | * for. Perhaps should add a new macro here.
|
---|
230 | *
|
---|
231 | * RED-PEN: vmalloc is not supported right now.
|
---|
232 | */
|
---|
233 | if (!pfn_valid(pfn))
|
---|
234 | return -EIO;
|
---|
235 |
|
---|
236 | vma->vm_pgoff = pfn;
|
---|
237 | return mmap_mem(file, vma);
|
---|
238 | }
|
---|
239 |
|
---|
240 | static int my_permission(struct inode *inode, int op)
|
---|
241 | {
|
---|
242 | /*
|
---|
243 | * only root (uid 0) may read from it
|
---|
244 | */
|
---|
245 | if (debug)
|
---|
246 | {
|
---|
247 | printk(KERN_INFO "samhain_kmem: permission op = %d, current->euid = %d\n",
|
---|
248 | op, (int)current->TASK_EUID );
|
---|
249 | }
|
---|
250 |
|
---|
251 | if ((op & 4) != 0 && (op & 2) == 0 && current->TASK_EUID == 0)
|
---|
252 | {
|
---|
253 | if (debug)
|
---|
254 | {
|
---|
255 | printk(KERN_INFO "samhain_kmem: access granted\n" );
|
---|
256 | }
|
---|
257 | return 0;
|
---|
258 | }
|
---|
259 |
|
---|
260 | /*
|
---|
261 | * If it's anything else, access is denied
|
---|
262 | */
|
---|
263 | if ((op & 2) != 0)
|
---|
264 | {
|
---|
265 | printk(KERN_INFO "/proc/kmem: access denied, "
|
---|
266 | "permission op = %d, current->euid = %d\n",
|
---|
267 | op, (int)current->TASK_EUID );
|
---|
268 | }
|
---|
269 | else if (debug)
|
---|
270 | {
|
---|
271 | printk(KERN_INFO "samhain_kmem: access denied\n" );
|
---|
272 | }
|
---|
273 | return -EACCES;
|
---|
274 | }
|
---|
275 |
|
---|
276 | static struct inode_operations Inode_Ops_Kmem = {
|
---|
277 | .permission = my_permission, /* check for permissions */
|
---|
278 | };
|
---|
279 |
|
---|
280 | static int open_kmem(struct inode * inode, struct file * filp)
|
---|
281 | {
|
---|
282 | int ret = capable(CAP_SYS_RAWIO) ? 0 : -EPERM;
|
---|
283 |
|
---|
284 | if (debug)
|
---|
285 | {
|
---|
286 | printk(KERN_INFO "samhain_kmem: open_kmem retval = %d\n", ret);
|
---|
287 | }
|
---|
288 |
|
---|
289 | if (ret == 0)
|
---|
290 | try_module_get(THIS_MODULE);
|
---|
291 |
|
---|
292 | if (debug)
|
---|
293 | {
|
---|
294 | printk(KERN_INFO "samhain_kmem: open_kmem return\n");
|
---|
295 | }
|
---|
296 |
|
---|
297 | return ret;
|
---|
298 | }
|
---|
299 |
|
---|
300 | static int close_kmem(struct inode *inode, struct file *file)
|
---|
301 | {
|
---|
302 | if (debug)
|
---|
303 | {
|
---|
304 | printk(KERN_INFO "samhain_kmem: close_kmem enter\n");
|
---|
305 | }
|
---|
306 |
|
---|
307 | module_put(THIS_MODULE);
|
---|
308 |
|
---|
309 | if (debug)
|
---|
310 | {
|
---|
311 | printk(KERN_INFO "samhain_kmem: close_kmem return\n");
|
---|
312 | }
|
---|
313 |
|
---|
314 | return 0; /* success */
|
---|
315 | }
|
---|
316 |
|
---|
317 | /*********************************************************************
|
---|
318 | *
|
---|
319 | * >>> Required info from System.map: vmlist_lock, vmlist <<<
|
---|
320 | */
|
---|
321 | static rwlock_t * sh_vmlist_lock_ptr = (rwlock_t *) SH_VMLIST_LOCK;
|
---|
322 |
|
---|
323 | static struct vm_struct * sh_vmlist = (struct vm_struct *) SH_VMLIST_LOCK;
|
---|
324 | /*
|
---|
325 | *
|
---|
326 | *********************************************************************/
|
---|
327 |
|
---|
328 | static long my_vread(char *buf, char *addr, unsigned long count)
|
---|
329 | {
|
---|
330 | struct vm_struct *tmp;
|
---|
331 | char *vaddr, *buf_start = buf;
|
---|
332 | unsigned long n;
|
---|
333 |
|
---|
334 | /* Don't allow overflow */
|
---|
335 | if ((unsigned long) addr + count < count)
|
---|
336 | count = -(unsigned long) addr;
|
---|
337 |
|
---|
338 | read_lock(sh_vmlist_lock_ptr);
|
---|
339 | for (tmp = sh_vmlist; tmp; tmp = tmp->next) {
|
---|
340 | vaddr = (char *) tmp->addr;
|
---|
341 | if (addr >= vaddr + tmp->size - PAGE_SIZE)
|
---|
342 | continue;
|
---|
343 | while (addr < vaddr) {
|
---|
344 | if (count == 0)
|
---|
345 | goto finished;
|
---|
346 | *buf = '\0';
|
---|
347 | buf++;
|
---|
348 | addr++;
|
---|
349 | count--;
|
---|
350 | }
|
---|
351 | n = vaddr + tmp->size - PAGE_SIZE - addr;
|
---|
352 | do {
|
---|
353 | if (count == 0)
|
---|
354 | goto finished;
|
---|
355 | *buf = *addr;
|
---|
356 | buf++;
|
---|
357 | addr++;
|
---|
358 | count--;
|
---|
359 | } while (--n > 0);
|
---|
360 | }
|
---|
361 | finished:
|
---|
362 | read_unlock(sh_vmlist_lock_ptr);
|
---|
363 | if (debug)
|
---|
364 | {
|
---|
365 | printk(KERN_INFO "samhain_kmem: start %lu\n", (unsigned long) buf_start);
|
---|
366 | printk(KERN_INFO "samhain_kmem: end %lu\n", (unsigned long) buf);
|
---|
367 | printk(KERN_INFO "samhain_kmem: size %lu\n", (unsigned long) (buf - buf_start));
|
---|
368 | }
|
---|
369 | return buf - buf_start;
|
---|
370 | }
|
---|
371 |
|
---|
372 | static ssize_t read_kmem(struct file *file, char __user *buf,
|
---|
373 | size_t count, loff_t *ppos)
|
---|
374 | {
|
---|
375 | unsigned long p = *ppos;
|
---|
376 | ssize_t low_count, read, sz;
|
---|
377 | char * kbuf; /* k-addr because vread() takes vmlist_lock rwlock */
|
---|
378 |
|
---|
379 | if (debug) {
|
---|
380 | printk(KERN_INFO "samhain_kmem: read_kmem entry\n");
|
---|
381 | printk(KERN_INFO "samhain_kmem: p %lu\n", (unsigned long) p);
|
---|
382 | printk(KERN_INFO "samhain_kmem: high %lu\n", (unsigned long) high_memory);
|
---|
383 | }
|
---|
384 |
|
---|
385 | read = 0;
|
---|
386 | if (p < (unsigned long) high_memory) {
|
---|
387 | low_count = count;
|
---|
388 |
|
---|
389 | if (debug) {
|
---|
390 | printk(KERN_INFO "samhain_kmem: low_count(1) %ld\n", (long) low_count);
|
---|
391 | }
|
---|
392 |
|
---|
393 | if (count > (unsigned long) high_memory - p)
|
---|
394 | low_count = (unsigned long) high_memory - p;
|
---|
395 |
|
---|
396 | if (debug) {
|
---|
397 | printk(KERN_INFO "samhain_kmem: low_count(2) %ld\n", (long) low_count);
|
---|
398 | }
|
---|
399 |
|
---|
400 | #ifdef __ARCH_HAS_NO_PAGE_ZERO_MAPPED
|
---|
401 | /* we don't have page 0 mapped on sparc and m68k.. */
|
---|
402 | if (p < PAGE_SIZE && low_count > 0) {
|
---|
403 | size_t tmp = PAGE_SIZE - p;
|
---|
404 | if (tmp > low_count) tmp = low_count;
|
---|
405 | if (clear_user(buf, tmp))
|
---|
406 | {
|
---|
407 | if (debug) {
|
---|
408 | printk(KERN_INFO "samhain_kmem: Bad address, line %d\n", __LINE__);
|
---|
409 | }
|
---|
410 | return -EFAULT;
|
---|
411 | }
|
---|
412 | buf += tmp;
|
---|
413 | p += tmp;
|
---|
414 | read += tmp;
|
---|
415 | low_count -= tmp;
|
---|
416 | count -= tmp;
|
---|
417 | }
|
---|
418 | #endif
|
---|
419 |
|
---|
420 | if (debug) {
|
---|
421 | printk(KERN_INFO "samhain_kmem: low_count(3) %ld\n", (long) low_count);
|
---|
422 | }
|
---|
423 |
|
---|
424 | while (low_count > 0) {
|
---|
425 | /*
|
---|
426 | * Handle first page in case it's not aligned
|
---|
427 | */
|
---|
428 | if (-p & (PAGE_SIZE - 1))
|
---|
429 | sz = -p & (PAGE_SIZE - 1);
|
---|
430 | else
|
---|
431 | sz = PAGE_SIZE;
|
---|
432 |
|
---|
433 | sz = min_t(unsigned long, sz, low_count);
|
---|
434 |
|
---|
435 | /*
|
---|
436 | * On ia64 if a page has been mapped somewhere as
|
---|
437 | * uncached, then it must also be accessed uncached
|
---|
438 | * by the kernel or data corruption may occur
|
---|
439 | */
|
---|
440 | kbuf = xlate_dev_kmem_ptr((char *)p);
|
---|
441 |
|
---|
442 | if (copy_to_user(buf, kbuf, sz))
|
---|
443 | {
|
---|
444 | if (debug) {
|
---|
445 | printk(KERN_INFO "samhain_kmem: Bad address, line %d\n", __LINE__);
|
---|
446 | printk(KERN_INFO "samhain_kmem: size %ld\n", (long) sz);
|
---|
447 | printk(KERN_INFO "samhain_kmem: kbuf %p\n", kbuf);
|
---|
448 | printk(KERN_INFO "samhain_kmem: buf %p\n", buf);
|
---|
449 | printk(KERN_INFO "samhain_kmem: high %lu\n", (unsigned long) high_memory);
|
---|
450 | }
|
---|
451 | return -EFAULT;
|
---|
452 | }
|
---|
453 | buf += sz;
|
---|
454 | p += sz;
|
---|
455 | read += sz;
|
---|
456 | low_count -= sz;
|
---|
457 | count -= sz;
|
---|
458 | if (debug) {
|
---|
459 | printk(KERN_INFO "samhain_kmem: low_count(4) %ld\n", (long) low_count);
|
---|
460 | }
|
---|
461 | }
|
---|
462 | }
|
---|
463 |
|
---|
464 | if (debug) {
|
---|
465 | printk(KERN_INFO "samhain_kmem: read_kmem mid\n");
|
---|
466 | printk(KERN_INFO "samhain_kmem: count %lu\n", (unsigned long) count);
|
---|
467 | }
|
---|
468 |
|
---|
469 | if (count > 0) {
|
---|
470 | kbuf = (char *)__get_free_page(GFP_KERNEL);
|
---|
471 | if (!kbuf)
|
---|
472 | {
|
---|
473 | if (debug) {
|
---|
474 | printk(KERN_INFO "samhain_kmem: out of memory\n");
|
---|
475 | }
|
---|
476 | return -ENOMEM;
|
---|
477 | }
|
---|
478 | while (count > 0) {
|
---|
479 | int len = count;
|
---|
480 |
|
---|
481 | if (len > PAGE_SIZE)
|
---|
482 | len = PAGE_SIZE;
|
---|
483 | len = my_vread(kbuf, (char *)p, len);
|
---|
484 | if (!len)
|
---|
485 | break;
|
---|
486 | if (copy_to_user(buf, kbuf, len)) {
|
---|
487 | if (debug) {
|
---|
488 | printk(KERN_INFO "samhain_kmem: Bad address, line %d\n", __LINE__);
|
---|
489 | printk(KERN_INFO "samhain_kmem: size %ld\n", (long) len);
|
---|
490 | printk(KERN_INFO "samhain_kmem: kbuf %p\n", kbuf);
|
---|
491 | printk(KERN_INFO "samhain_kmem: buf %p\n", buf);
|
---|
492 | printk(KERN_INFO "samhain_kmem: high %lu\n", (unsigned long) high_memory);
|
---|
493 | }
|
---|
494 | free_page((unsigned long)kbuf);
|
---|
495 | return -EFAULT;
|
---|
496 | }
|
---|
497 | count -= len;
|
---|
498 | buf += len;
|
---|
499 | read += len;
|
---|
500 | p += len;
|
---|
501 | }
|
---|
502 | free_page((unsigned long)kbuf);
|
---|
503 | }
|
---|
504 | *ppos = p;
|
---|
505 | if (debug) {
|
---|
506 | printk(KERN_INFO "samhain_kmem: read_kmem end\n");
|
---|
507 | printk(KERN_INFO "samhain_kmem: read %ld\n", (long) read);
|
---|
508 | }
|
---|
509 | return read;
|
---|
510 | }
|
---|
511 |
|
---|
512 |
|
---|
513 | static loff_t memory_lseek(struct file * file, loff_t offset, int orig)
|
---|
514 | {
|
---|
515 | loff_t ret;
|
---|
516 |
|
---|
517 | mutex_lock(&file->f_path.dentry->d_inode->i_mutex);
|
---|
518 | switch (orig) {
|
---|
519 | case 0:
|
---|
520 | file->f_pos = offset;
|
---|
521 | ret = file->f_pos;
|
---|
522 | force_successful_syscall_return();
|
---|
523 | break;
|
---|
524 | case 1:
|
---|
525 | file->f_pos += offset;
|
---|
526 | ret = file->f_pos;
|
---|
527 | force_successful_syscall_return();
|
---|
528 | break;
|
---|
529 | default:
|
---|
530 | if (debug) {
|
---|
531 | printk(KERN_INFO "samhain_kmem: invalid input %d\n", orig);
|
---|
532 | }
|
---|
533 | ret = -EINVAL;
|
---|
534 | }
|
---|
535 | mutex_unlock(&file->f_path.dentry->d_inode->i_mutex);
|
---|
536 | return ret;
|
---|
537 | }
|
---|
538 |
|
---|
539 | static const struct file_operations File_Ops_Kmem = {
|
---|
540 | .llseek = memory_lseek,
|
---|
541 | .read = read_kmem,
|
---|
542 | .mmap = mmap_kmem,
|
---|
543 | .open = open_kmem,
|
---|
544 | .release = close_kmem,
|
---|
545 | .get_unmapped_area = get_unmapped_area_mem,
|
---|
546 | };
|
---|
547 |
|
---|
548 |
|
---|
549 | /* Init function called on module entry
|
---|
550 | */
|
---|
551 | static int my_module_init( void )
|
---|
552 | {
|
---|
553 | int ret = 0;
|
---|
554 |
|
---|
555 | proc_entry = create_proc_entry( "kmem", 0400, NULL );
|
---|
556 |
|
---|
557 | if (proc_entry == NULL) {
|
---|
558 |
|
---|
559 | ret = -ENOMEM;
|
---|
560 |
|
---|
561 | printk(KERN_INFO "samhain_kmem: Couldn't create proc entry\n");
|
---|
562 |
|
---|
563 | } else {
|
---|
564 |
|
---|
565 | /* 2.6.30 */
|
---|
566 | #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,30)
|
---|
567 | proc_entry->owner = THIS_MODULE;
|
---|
568 | #endif
|
---|
569 | proc_entry->proc_iops = &Inode_Ops_Kmem;
|
---|
570 | proc_entry->proc_fops = &File_Ops_Kmem;
|
---|
571 |
|
---|
572 | proc_entry->uid = 0;
|
---|
573 | proc_entry->gid = 0;
|
---|
574 | proc_entry->mode = S_IFREG | S_IRUSR;
|
---|
575 |
|
---|
576 | if (debug) {
|
---|
577 | printk(KERN_INFO "samhain_kmem: module is now loaded.\n");
|
---|
578 | }
|
---|
579 | }
|
---|
580 |
|
---|
581 | return ret;
|
---|
582 | }
|
---|
583 |
|
---|
584 | /* Cleanup function called on module exit */
|
---|
585 |
|
---|
586 | static void my_module_cleanup( void )
|
---|
587 | {
|
---|
588 | remove_proc_entry("kmem", NULL);
|
---|
589 |
|
---|
590 | if (debug) {
|
---|
591 | printk(KERN_INFO "samhain_kmem: module is now unloaded.\n");
|
---|
592 | }
|
---|
593 | return;
|
---|
594 | }
|
---|
595 |
|
---|
596 |
|
---|
597 |
|
---|
598 | /* Declare entry and exit functions */
|
---|
599 |
|
---|
600 | module_init( my_module_init );
|
---|
601 |
|
---|
602 | module_exit( my_module_cleanup );
|
---|