| 1 | /* $Id: memory.cpp,v 1.1.1.1 2003/07/02 13:57:02 eleph Exp $ */
|
|---|
| 2 | /*
|
|---|
| 3 | * OS/2 implementation of Linux memory kernel services
|
|---|
| 4 | *
|
|---|
| 5 | * (C) 2000-2002 InnoTek Systemberatung GmbH
|
|---|
| 6 | * (C) 2000-2001 Sander van Leeuwen (sandervl@xs4all.nl)
|
|---|
| 7 | *
|
|---|
| 8 | * This program is free software; you can redistribute it and/or
|
|---|
| 9 | * modify it under the terms of the GNU General Public License as
|
|---|
| 10 | * published by the Free Software Foundation; either version 2 of
|
|---|
| 11 | * the License, or (at your option) any later version.
|
|---|
| 12 | *
|
|---|
| 13 | * This program is distributed in the hope that it will be useful,
|
|---|
| 14 | * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|---|
| 15 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|---|
| 16 | * GNU General Public License for more details.
|
|---|
| 17 | *
|
|---|
| 18 | * You should have received a copy of the GNU General Public
|
|---|
| 19 | * License along with this program; if not, write to the Free
|
|---|
| 20 | * Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139,
|
|---|
| 21 | * USA.
|
|---|
| 22 | *
|
|---|
| 23 | */
|
|---|
| 24 |
|
|---|
| 25 | #define INCL_NOPMAPI
|
|---|
| 26 | #define INCL_DOSERRORS // for ERROR_INVALID_FUNCTION
|
|---|
| 27 | #include <os2.h>
|
|---|
| 28 | #include <devhelp.h>
|
|---|
| 29 | #include <ossidc.h>
|
|---|
| 30 | #include <string.h>
|
|---|
| 31 | #include <dbgos2.h>
|
|---|
| 32 | #include <stacktoflat.h>
|
|---|
| 33 | #include <limits.h>
|
|---|
| 34 | #include <kee.h>
|
|---|
| 35 | #include "malloc.h"
|
|---|
| 36 | #define _I386_PAGE_H
|
|---|
| 37 | typedef struct { unsigned long pgprot; } pgprot_t;
|
|---|
| 38 | #define MAP_NR(addr) (__pa(addr) >> PAGE_SHIFT)
|
|---|
| 39 | #define PAGE_SHIFT 12
|
|---|
| 40 | #define __PAGE_OFFSET (0xC0000000)
|
|---|
| 41 |
|
|---|
| 42 | #define PAGE_OFFSET ((unsigned long)__PAGE_OFFSET)
|
|---|
| 43 | #define __pa(x) ((unsigned long)(x)-PAGE_OFFSET)
|
|---|
| 44 |
|
|---|
| 45 | #include <linux/mm.h>
|
|---|
| 46 | #include <linux/slab.h>
|
|---|
| 47 | #include <linux/printk.h>
|
|---|
| 48 |
|
|---|
| 49 | #pragma off (unreferenced)
|
|---|
| 50 |
|
|---|
| 51 | #define PAGE_SIZE 4096
|
|---|
| 52 | #define min(a,b) (((a) < (b)) ? (a) : (b))
|
|---|
| 53 |
|
|---|
| 54 | int free_pages(unsigned long addr, unsigned long order);
|
|---|
| 55 | int __compat_get_order(unsigned long size);
|
|---|
| 56 | #ifdef DEBUGHEAP
|
|---|
| 57 | void near *__kmalloc(int size, int flags, const char *filename, int lineno);
|
|---|
| 58 | void __kfree(const void near *ptr, const char *filename, int lineno);
|
|---|
| 59 | #else
|
|---|
| 60 | void near *__kmalloc(int size, int flags);
|
|---|
| 61 | void __kfree(const void near *ptr);
|
|---|
| 62 | #endif
|
|---|
| 63 |
|
|---|
| 64 | #ifdef DEBUGHEAP
|
|---|
| 65 | #define _kmalloc(a, b) __kmalloc(a, b, __FILE__, __LINE__)
|
|---|
| 66 | #define _kfree(a) __kfree(a, __FILE__, __LINE__)
|
|---|
| 67 | #else
|
|---|
| 68 | #define _kmalloc(a, b) __kmalloc(a, b)
|
|---|
| 69 | #define _kfree(a) __kfree(a)
|
|---|
| 70 | #endif
|
|---|
| 71 |
|
|---|
| 72 | typedef struct _BaseAddr {
|
|---|
| 73 | ULONG base; // VMAlloc addr
|
|---|
| 74 | ULONG retaddr; // aligned addr returned to caller
|
|---|
| 75 | ULONG size; // VMAlloc size
|
|---|
| 76 | struct _BaseAddr NEAR *next;
|
|---|
| 77 | } BaseAddr;
|
|---|
| 78 |
|
|---|
| 79 | static BaseAddr NEAR *pBaseAddrHead = NULL;
|
|---|
| 80 |
|
|---|
| 81 | //******************************************************************************
|
|---|
| 82 | //Very simple linked list for storing original addresses returned by VMAlloc
|
|---|
| 83 | //if returned address is different due to alignment requirements
|
|---|
| 84 | //Performance is not an issue as the alloc & free functions aren't called often.
|
|---|
| 85 | //(e.g. ALS4000 driver calls it 4 times (2 alloc during boot, 2 during shutdown)
|
|---|
| 86 | //******************************************************************************
|
|---|
| 87 | void AddBaseAddress(ULONG baseaddr, ULONG retaddr, ULONG size)
|
|---|
| 88 | {
|
|---|
| 89 | BaseAddr NEAR *pBase;
|
|---|
| 90 |
|
|---|
| 91 | pBase = (BaseAddr NEAR *)_kmalloc(sizeof(BaseAddr), 0);
|
|---|
| 92 | if(pBase == NULL) {
|
|---|
| 93 | DebugInt3();
|
|---|
| 94 | return;
|
|---|
| 95 | }
|
|---|
| 96 | DevCli();
|
|---|
| 97 | pBase->base = baseaddr;
|
|---|
| 98 | pBase->retaddr = retaddr;
|
|---|
| 99 | pBase->size = size;
|
|---|
| 100 | pBase->next = pBaseAddrHead;
|
|---|
| 101 | pBaseAddrHead = pBase;
|
|---|
| 102 | DevSti();
|
|---|
| 103 | }
|
|---|
| 104 | //******************************************************************************
|
|---|
| 105 | //******************************************************************************
|
|---|
| 106 | ULONG GetBaseAddressAndFree(ULONG addr, ULONG *pSize)
|
|---|
| 107 | {
|
|---|
| 108 | BaseAddr NEAR *pCur, NEAR *pTemp;
|
|---|
| 109 |
|
|---|
| 110 | if(pBaseAddrHead == NULL) return addr;
|
|---|
| 111 |
|
|---|
| 112 | DevCli();
|
|---|
| 113 | pCur = pBaseAddrHead;
|
|---|
| 114 |
|
|---|
| 115 | // If address is in list, remove list item and free entry
|
|---|
| 116 | // Caller must VMFree returned address or else
|
|---|
| 117 | if(pCur->retaddr == addr)
|
|---|
| 118 | {
|
|---|
| 119 | addr = pCur->base;
|
|---|
| 120 | if(pSize) *pSize = pCur->size;
|
|---|
| 121 | pBaseAddrHead = pCur->next;
|
|---|
| 122 | _kfree(pCur);
|
|---|
| 123 | }
|
|---|
| 124 | else
|
|---|
| 125 | while(pCur->next) {
|
|---|
| 126 | if(pCur->next->retaddr == addr) {
|
|---|
| 127 | pTemp = pCur->next;
|
|---|
| 128 | addr = pTemp->base;
|
|---|
| 129 | if(pSize) *pSize = pTemp->size;
|
|---|
| 130 | pCur->next = pTemp->next;
|
|---|
| 131 |
|
|---|
| 132 | _kfree(pTemp);
|
|---|
| 133 | break;
|
|---|
| 134 | }
|
|---|
| 135 | pCur = pCur->next;
|
|---|
| 136 | }
|
|---|
| 137 | DevSti();
|
|---|
| 138 | return addr;
|
|---|
| 139 | }
|
|---|
| 140 | //******************************************************************************
|
|---|
| 141 | //******************************************************************************
|
|---|
| 142 | ULONG GetBaseAddressNoFree(ULONG addr, ULONG *pSize)
|
|---|
| 143 | {
|
|---|
| 144 | BaseAddr NEAR *pCur, NEAR *pTemp;
|
|---|
| 145 |
|
|---|
| 146 | if(pBaseAddrHead == NULL) return addr;
|
|---|
| 147 |
|
|---|
| 148 | DevCli();
|
|---|
| 149 | pCur = pBaseAddrHead;
|
|---|
| 150 |
|
|---|
| 151 | if(pCur->retaddr == addr)
|
|---|
| 152 | {
|
|---|
| 153 | addr = pCur->base;
|
|---|
| 154 | if(pSize) *pSize = pCur->size;
|
|---|
| 155 | pBaseAddrHead = pCur->next;
|
|---|
| 156 | // _kfree(pCur);
|
|---|
| 157 | }
|
|---|
| 158 | else
|
|---|
| 159 | while(pCur->next) {
|
|---|
| 160 | if(pCur->next->retaddr == addr) {
|
|---|
| 161 | pTemp = pCur->next;
|
|---|
| 162 | addr = pTemp->base;
|
|---|
| 163 | if(pSize) *pSize = pTemp->size;
|
|---|
| 164 | pCur->next = pTemp->next;
|
|---|
| 165 | // _kfree(pTemp);
|
|---|
| 166 | break;
|
|---|
| 167 | }
|
|---|
| 168 | pCur = pCur->next;
|
|---|
| 169 | }
|
|---|
| 170 | DevSti();
|
|---|
| 171 | return addr;
|
|---|
| 172 | }
|
|---|
| 173 | //******************************************************************************
|
|---|
| 174 | //NOTE: Assumes memory is continuous!!
|
|---|
| 175 | //******************************************************************************
|
|---|
| 176 | unsigned long virt_to_phys(void * address)
|
|---|
| 177 | {
|
|---|
| 178 | KEEVMPageList pagelist;
|
|---|
| 179 | ULONG nrpages;
|
|---|
| 180 |
|
|---|
| 181 | if(KernLinToPageList(address, PAGE_SIZE, &pagelist, &nrpages)) {
|
|---|
| 182 | DebugInt3();
|
|---|
| 183 | return 0;
|
|---|
| 184 | }
|
|---|
| 185 | return pagelist.addr;
|
|---|
| 186 | }
|
|---|
| 187 | //******************************************************************************
|
|---|
| 188 | //******************************************************************************
|
|---|
| 189 | void * phys_to_virt(unsigned long address)
|
|---|
| 190 | {
|
|---|
| 191 | APIRET rc = 0;
|
|---|
| 192 | ULONG addr = 0;
|
|---|
| 193 |
|
|---|
| 194 | SHORT sel;
|
|---|
| 195 | rc = KernVMAlloc(PAGE_SIZE, VMDHA_PHYS, (PVOID*)&addr, (PVOID*)&address, &sel);
|
|---|
| 196 | if (rc != 0) {
|
|---|
| 197 | DebugInt3();
|
|---|
| 198 | return NULL;
|
|---|
| 199 | }
|
|---|
| 200 | return (void *)addr;
|
|---|
| 201 | }
|
|---|
| 202 | //******************************************************************************
|
|---|
| 203 | extern int fStrategyInit;
|
|---|
| 204 | //******************************************************************************
|
|---|
| 205 | APIRET VMAlloc(ULONG size, ULONG flags, LINEAR *pAddr)
|
|---|
| 206 | {
|
|---|
| 207 | APIRET rc;
|
|---|
| 208 | ULONG addr;
|
|---|
| 209 | SHORT sel;
|
|---|
| 210 |
|
|---|
| 211 | if(fStrategyInit && !(flags & VMDHA_16M)) {
|
|---|
| 212 | flags |= VMDHA_USEHIGHMEM;
|
|---|
| 213 | }
|
|---|
| 214 |
|
|---|
| 215 | __again:
|
|---|
| 216 |
|
|---|
| 217 | rc = KernVMAlloc(size, flags, (PVOID*)&addr, (PVOID*)-1, &sel);
|
|---|
| 218 | if (rc == 0) {
|
|---|
| 219 | *pAddr = (LINEAR)addr;
|
|---|
| 220 | if (flags & VMDHA_USEHIGHMEM)
|
|---|
| 221 | dprintf1((("allocated %X in HIGH memory\n"), size));
|
|---|
| 222 | else dprintf1((("allocated %X in LOW memory\n"), size));
|
|---|
| 223 | }
|
|---|
| 224 | if ((rc == 87) && (flags & VMDHA_USEHIGHMEM))
|
|---|
| 225 | {
|
|---|
| 226 | // EARLYMEMINIT workaround
|
|---|
| 227 | flags = flags & (~VMDHA_USEHIGHMEM);
|
|---|
| 228 | goto __again;
|
|---|
| 229 | }
|
|---|
| 230 | return rc;
|
|---|
| 231 | }
|
|---|
| 232 | //******************************************************************************
|
|---|
| 233 | //******************************************************************************
|
|---|
| 234 | APIRET VMFree(LINEAR addr)
|
|---|
| 235 | {
|
|---|
| 236 | APIRET rc;
|
|---|
| 237 | rc = KernVMFree((PVOID)addr);
|
|---|
| 238 | if(rc) {
|
|---|
| 239 | DebugInt3();
|
|---|
| 240 | }
|
|---|
| 241 |
|
|---|
| 242 | return rc;
|
|---|
| 243 | }
|
|---|
| 244 | //******************************************************************************
|
|---|
| 245 | ULONG ulget_free_pagesMemUsed = 0;
|
|---|
| 246 |
|
|---|
| 247 | //******************************************************************************
|
|---|
| 248 | //******************************************************************************
|
|---|
| 249 | void *__get_free_dma_pages(unsigned long size, unsigned long flags)
|
|---|
| 250 | {
|
|---|
| 251 | ULONG addr, physaddr, physaddr2, diff, tempaddr;
|
|---|
| 252 | APIRET rc;
|
|---|
| 253 |
|
|---|
| 254 | if(VMAlloc(size, flags, (LINEAR *)&addr)) {
|
|---|
| 255 | DebugInt3();
|
|---|
| 256 | return 0;
|
|---|
| 257 | }
|
|---|
| 258 | physaddr = virt_to_phys((void *)addr);
|
|---|
| 259 | if(physaddr) {
|
|---|
| 260 | ULONG startpage = (physaddr >> 16);
|
|---|
| 261 | ULONG endpage = (physaddr + ((size < 0x10000) ? size : 63*1024)) >> 16;
|
|---|
| 262 |
|
|---|
| 263 | if(startpage != endpage) {
|
|---|
| 264 | // not in same 32K page, try once more
|
|---|
| 265 | rc = VMAlloc(size, flags, (LINEAR *)&tempaddr);
|
|---|
| 266 | VMFree((LINEAR)addr);
|
|---|
| 267 | if(rc) {
|
|---|
| 268 | DebugInt3();
|
|---|
| 269 | return 0;
|
|---|
| 270 | }
|
|---|
| 271 | addr = tempaddr;
|
|---|
| 272 |
|
|---|
| 273 | physaddr = virt_to_phys((void *)addr);
|
|---|
| 274 | if(physaddr) {
|
|---|
| 275 | ULONG startpage = (physaddr >> 16);
|
|---|
| 276 | ULONG endpage = (physaddr + ((size < 0x10000) ? size : 63*1024)) >> 16;
|
|---|
| 277 |
|
|---|
| 278 | if(startpage != endpage) {
|
|---|
| 279 | //oops, this didn't work, fail
|
|---|
| 280 | VMFree((LINEAR)addr);
|
|---|
| 281 | dprintf(("get_free_dma_pages failed %x size:%x st:%x end:%x, trying wasteful method instead", physaddr, size, startpage, endpage));
|
|---|
| 282 | return 0;
|
|---|
| 283 | }
|
|---|
| 284 | }
|
|---|
| 285 | }
|
|---|
| 286 | }
|
|---|
| 287 | else {
|
|---|
| 288 | DebugInt3();
|
|---|
| 289 | VMFree((LINEAR)addr);
|
|---|
| 290 | addr = 0;
|
|---|
| 291 | }
|
|---|
| 292 |
|
|---|
| 293 | if(addr) {
|
|---|
| 294 | //only done to save size of memory block
|
|---|
| 295 | AddBaseAddress(addr, addr, size);
|
|---|
| 296 | ulget_free_pagesMemUsed += size;
|
|---|
| 297 | dprintf(("get_free_dma_pages: size=%x adr=%x (phys %x) total alloc size=%x",
|
|---|
| 298 | size, (ULONG)addr, virt_to_phys((void *)addr), ulget_free_pagesMemUsed));
|
|---|
| 299 | }
|
|---|
| 300 |
|
|---|
| 301 | return (void *)addr;
|
|---|
| 302 | }
|
|---|
| 303 | //******************************************************************************
|
|---|
| 304 | //******************************************************************************
|
|---|
| 305 | void *__get_free_pages(int gfp_mask, unsigned long order)
|
|---|
| 306 | {
|
|---|
| 307 | ULONG addr;
|
|---|
| 308 | ULONG flags = VMDHA_FIXED|VMDHA_CONTIG;
|
|---|
| 309 | ULONG size, allocsize;
|
|---|
| 310 |
|
|---|
| 311 | order = (1 << order); //TODO: Is this correct???
|
|---|
| 312 | size = order * PAGE_SIZE;
|
|---|
| 313 |
|
|---|
| 314 | if(gfp_mask & (GFP_DMA|GFP_DMAHIGHMEM))
|
|---|
| 315 | {//below 16 mb for legacy DMA?
|
|---|
| 316 | if(gfp_mask & GFP_DMA)
|
|---|
| 317 | flags |= VMDHA_16M;
|
|---|
| 318 |
|
|---|
| 319 | //these buffers must be aligned at 64kb boundary
|
|---|
| 320 |
|
|---|
| 321 | //first try a less wasteful approach
|
|---|
| 322 | void *pBlock;
|
|---|
| 323 |
|
|---|
| 324 | pBlock = __get_free_dma_pages(size, flags);
|
|---|
| 325 | if(pBlock) {
|
|---|
| 326 | return pBlock;
|
|---|
| 327 | }
|
|---|
| 328 | //else allocate extra memory to make sure we can satisfy
|
|---|
| 329 | //the alignment requirement
|
|---|
| 330 | if(size < 0x10000) {
|
|---|
| 331 | allocsize = size * 2;
|
|---|
| 332 | }
|
|---|
| 333 | else {
|
|---|
| 334 | allocsize = size + 0x10000;
|
|---|
| 335 | }
|
|---|
| 336 | }
|
|---|
| 337 | else allocsize = size;
|
|---|
| 338 |
|
|---|
| 339 | if(VMAlloc(allocsize, flags, (LINEAR *)&addr)) {
|
|---|
| 340 | DebugInt3();
|
|---|
| 341 | return 0;
|
|---|
| 342 | }
|
|---|
| 343 | //dprintf(("__get_free_pages %d returned %x", order*PAGE_SIZE, addr));
|
|---|
| 344 | if(gfp_mask & (GFP_DMA|GFP_DMAHIGHMEM))
|
|---|
| 345 | {//must be aligned at 64kb boundary
|
|---|
| 346 | ULONG physaddr = virt_to_phys((void *)addr);
|
|---|
| 347 | ULONG physaddr2;
|
|---|
| 348 |
|
|---|
| 349 | if(physaddr) {
|
|---|
| 350 | ULONG startpage = (physaddr >> 16);
|
|---|
| 351 | ULONG endpage = (physaddr + ((size < 0x10000) ? size : 63*1024)) >> 16;
|
|---|
| 352 |
|
|---|
| 353 | if (startpage != endpage) {
|
|---|
| 354 | // Not in same 32K page
|
|---|
| 355 | physaddr2 = (startpage+1) << 16;
|
|---|
| 356 |
|
|---|
| 357 | AddBaseAddress(addr, addr + (physaddr2 - physaddr), allocsize);
|
|---|
| 358 | addr += (physaddr2 - physaddr);
|
|---|
| 359 | }
|
|---|
| 360 | }
|
|---|
| 361 | else {
|
|---|
| 362 | DebugInt3();
|
|---|
| 363 | free_pages(addr, order);
|
|---|
| 364 | addr = 0;
|
|---|
| 365 | }
|
|---|
| 366 | }
|
|---|
| 367 | else {
|
|---|
| 368 | //only done to save size of memory block
|
|---|
| 369 | AddBaseAddress(addr, addr, allocsize);
|
|---|
| 370 | }
|
|---|
| 371 | if(addr) {
|
|---|
| 372 | //dprintf(("get_free_pages %d (%d) -> %x (phys %x)", allocsize, size, (ULONG)addr, virt_to_phys((void *)addr)));
|
|---|
| 373 | ulget_free_pagesMemUsed += allocsize;
|
|---|
| 374 | //dprintf(("get_free_pages: total alloc size %d", ulget_free_pagesMemUsed));
|
|---|
| 375 | }
|
|---|
| 376 | return (void *)addr;
|
|---|
| 377 | }
|
|---|
| 378 | //******************************************************************************
|
|---|
| 379 | //******************************************************************************
|
|---|
| 380 | int free_pages(unsigned long addr, unsigned long order)
|
|---|
| 381 | {
|
|---|
| 382 | ULONG rc, size = 0;
|
|---|
| 383 |
|
|---|
| 384 | //check if it really is the base of the allocation (see above)
|
|---|
| 385 | addr = GetBaseAddressAndFree(addr, (ULONG NEAR *)&size);
|
|---|
| 386 |
|
|---|
| 387 | if(VMFree((LINEAR)addr)) {
|
|---|
| 388 | DebugInt3();
|
|---|
| 389 | }
|
|---|
| 390 | else {
|
|---|
| 391 | //dprintf(("free_pages %x size %d", (ULONG)addr, size));
|
|---|
| 392 | ulget_free_pagesMemUsed -= size;
|
|---|
| 393 | //dprintf(("free_pages: total alloc size %d", ulget_free_pagesMemUsed));
|
|---|
| 394 | }
|
|---|
| 395 | //dprintf(("free_pages %x", addr));
|
|---|
| 396 | return 0;
|
|---|
| 397 | }
|
|---|
| 398 | //******************************************************************************
|
|---|
| 399 | //******************************************************************************
|
|---|
| 400 | void *vmalloc(unsigned long size)
|
|---|
| 401 | {
|
|---|
| 402 | ULONG addr = 0;
|
|---|
| 403 | ULONG flags = VMDHA_FIXED|VMDHA_CONTIG;
|
|---|
| 404 |
|
|---|
| 405 | //round to next page boundary
|
|---|
| 406 | size = size + PAGE_SIZE - 1;
|
|---|
| 407 | size &= 0xFFFFF000;
|
|---|
| 408 |
|
|---|
| 409 | if(VMAlloc(size, flags, (LINEAR *)&addr)) {
|
|---|
| 410 | DebugInt3();
|
|---|
| 411 | return 0;
|
|---|
| 412 | }
|
|---|
| 413 | if(addr) {
|
|---|
| 414 | //dprintf(("vmalloc %d -> %x (phys %x)", size, addr, virt_to_phys((void *)addr)));
|
|---|
| 415 | //only done to save size of memory block
|
|---|
| 416 | AddBaseAddress(addr, addr, size);
|
|---|
| 417 | ulget_free_pagesMemUsed += size;
|
|---|
| 418 | //dprintf(("vmalloc: total alloc size %d", ulget_free_pagesMemUsed));
|
|---|
| 419 | }
|
|---|
| 420 | return (void *)addr;
|
|---|
| 421 | }
|
|---|
| 422 | //******************************************************************************
|
|---|
| 423 | //******************************************************************************
|
|---|
| 424 | void *__vmalloc(unsigned long size, gfp_t gfp_mask)
|
|---|
| 425 | {
|
|---|
| 426 | return vmalloc(size);
|
|---|
| 427 | }
|
|---|
| 428 | //******************************************************************************
|
|---|
| 429 | //******************************************************************************
|
|---|
| 430 | /**
|
|---|
| 431 | * __vmalloc_node - allocate virtually contiguous memory
|
|---|
| 432 | * @size: allocation size
|
|---|
| 433 | * @align: desired alignment
|
|---|
| 434 | * @gfp_mask: flags for the page level allocator
|
|---|
| 435 | * @node: node to use for allocation or NUMA_NO_NODE
|
|---|
| 436 | * @caller: caller's return address
|
|---|
| 437 | *
|
|---|
| 438 | * Allocate enough pages to cover @size from the page level allocator with
|
|---|
| 439 | * @gfp_mask flags. Map them into contiguous kernel virtual space.
|
|---|
| 440 | *
|
|---|
| 441 | * Reclaim modifiers in @gfp_mask - __GFP_NORETRY, __GFP_RETRY_MAYFAIL
|
|---|
| 442 | * and __GFP_NOFAIL are not supported
|
|---|
| 443 | *
|
|---|
| 444 | * Any use of gfp flags outside of GFP_KERNEL should be consulted
|
|---|
| 445 | * with mm people.
|
|---|
| 446 | *
|
|---|
| 447 | * Return: pointer to the allocated memory or %NULL on error
|
|---|
| 448 | */
|
|---|
| 449 | void *__vmalloc_node(unsigned long size, unsigned long align,
|
|---|
| 450 | gfp_t gfp_mask, int node, const void *caller)
|
|---|
| 451 | {
|
|---|
| 452 | return vmalloc(size);
|
|---|
| 453 | }
|
|---|
| 454 | //******************************************************************************
|
|---|
| 455 | //******************************************************************************
|
|---|
| 456 | void vfree(void *ptr)
|
|---|
| 457 | {
|
|---|
| 458 | APIRET rc;
|
|---|
| 459 | ULONG size = 0;
|
|---|
| 460 |
|
|---|
| 461 | GetBaseAddressAndFree((ULONG)ptr, (ULONG NEAR *)&size);
|
|---|
| 462 |
|
|---|
| 463 | if(VMFree((LINEAR)ptr)) {
|
|---|
| 464 | DebugInt3();
|
|---|
| 465 | }
|
|---|
| 466 | else {
|
|---|
| 467 | //dprintf(("vfree %x size %d", (ULONG)ptr, size));
|
|---|
| 468 | ulget_free_pagesMemUsed -= size;
|
|---|
| 469 | //dprintf(("vfree: total alloc size %d", ulget_free_pagesMemUsed));
|
|---|
| 470 | }
|
|---|
| 471 | }
|
|---|
| 472 | //******************************************************************************
|
|---|
| 473 | //******************************************************************************
|
|---|
| 474 | struct page * alloc_pages(int gfp_mask, unsigned long order)
|
|---|
| 475 | {
|
|---|
| 476 | DebugInt3();
|
|---|
| 477 | return 0;
|
|---|
| 478 | }
|
|---|
| 479 | //******************************************************************************
|
|---|
| 480 | //******************************************************************************
|
|---|
| 481 | int remap_page_range(unsigned long from, unsigned long to, unsigned long size, pgprot_t prot)
|
|---|
| 482 | {
|
|---|
| 483 | DebugInt3();
|
|---|
| 484 | return 0;
|
|---|
| 485 | }
|
|---|
| 486 | //******************************************************************************
|
|---|
| 487 | //Map physical address (memory mapped io range) to linear
|
|---|
| 488 | //******************************************************************************
|
|---|
| 489 | void * __ioremap(unsigned long physaddr, unsigned long size, unsigned long flags)
|
|---|
| 490 | {
|
|---|
| 491 | ULONG addr = 0, Offset = 0, PhysicalAddress = 0, Length = 0;
|
|---|
| 492 | APIRET rc;
|
|---|
| 493 |
|
|---|
| 494 | PhysicalAddress = physaddr;
|
|---|
| 495 | Length = size;
|
|---|
| 496 |
|
|---|
| 497 | Offset = PhysicalAddress & (PAGE_SIZE - 1); // within Page
|
|---|
| 498 | Length = (Length + Offset + (PAGE_SIZE - 1)) & ~(PAGE_SIZE - 1);
|
|---|
| 499 | PhysicalAddress &= ~(PAGE_SIZE - 1);
|
|---|
| 500 | //dprintf(("ioremap: len %d phys %x off %x", Length, PhysicalAddress, Offset));
|
|---|
| 501 |
|
|---|
| 502 | //round to next page boundary
|
|---|
| 503 | //size = size + PAGE_SIZE - 1;
|
|---|
| 504 | //size &= 0xFFFFF000;
|
|---|
| 505 |
|
|---|
| 506 | SHORT sel;
|
|---|
| 507 | //rc = KernVMAlloc(size, VMDHA_PHYS, (PVOID*)&addr, (PVOID*)&physaddr, &sel);
|
|---|
| 508 | rc = KernVMAlloc(Length, VMDHA_PHYS, (PVOID*)&addr, (PVOID*)&PhysicalAddress, &sel);
|
|---|
| 509 | if (rc != 0) {
|
|---|
| 510 | dprintf(("ioremap error: %x", rc));
|
|---|
| 511 | DebugInt3();
|
|---|
| 512 | return NULL;
|
|---|
| 513 | }
|
|---|
| 514 | return (void *)( addr + Offset) ; //PS
|
|---|
| 515 | }
|
|---|
| 516 | //******************************************************************************
|
|---|
| 517 | //******************************************************************************
|
|---|
| 518 | void iounmap(void *addr)
|
|---|
| 519 | {
|
|---|
| 520 | LINEAR ad;
|
|---|
| 521 |
|
|---|
| 522 | ad=(LINEAR)addr;
|
|---|
| 523 | // *ad &= ~(0xfff); // 12 Jun 07 SHL this looks wrong
|
|---|
| 524 | ad = (LINEAR)((ULONG)ad & ~0xfff); // 12 Jun 07 SHL Round down to 4KiB
|
|---|
| 525 | if(VMFree((LINEAR)ad)) {
|
|---|
| 526 | DebugInt3();
|
|---|
| 527 | }
|
|---|
| 528 | }
|
|---|
| 529 | //******************************************************************************
|
|---|
| 530 | //******************************************************************************
|
|---|
| 531 | int is_access_ok(int type, void *addr, unsigned long size)
|
|---|
| 532 | {
|
|---|
| 533 | return 1;
|
|---|
| 534 | }
|
|---|
| 535 | //******************************************************************************
|
|---|
| 536 | //******************************************************************************
|
|---|
| 537 | void __copy_user(void *to, const void *from, unsigned long n)
|
|---|
| 538 | {
|
|---|
| 539 | if(to == NULL || from == NULL) {
|
|---|
| 540 | DebugInt3();
|
|---|
| 541 | return;
|
|---|
| 542 | }
|
|---|
| 543 | if(n == 0) return;
|
|---|
| 544 |
|
|---|
| 545 | memcpy(to, from, n);
|
|---|
| 546 | }
|
|---|
| 547 | //******************************************************************************
|
|---|
| 548 | //******************************************************************************
|
|---|
| 549 | unsigned long copy_to_user(void *to, const void *from, unsigned long n)
|
|---|
| 550 | {
|
|---|
| 551 | if(to == NULL || from == NULL) {
|
|---|
| 552 | DebugInt3();
|
|---|
| 553 | return 0;
|
|---|
| 554 | }
|
|---|
| 555 | if(n == 0) return 0;
|
|---|
| 556 |
|
|---|
| 557 | memcpy(to, from, n);
|
|---|
| 558 | return 0;
|
|---|
| 559 | }
|
|---|
| 560 | //******************************************************************************
|
|---|
| 561 | //******************************************************************************
|
|---|
| 562 | void __copy_user_zeroing(void *to, const void *from, unsigned long n)
|
|---|
| 563 | {
|
|---|
| 564 | if(to == NULL || from == NULL) {
|
|---|
| 565 | DebugInt3();
|
|---|
| 566 | return;
|
|---|
| 567 | }
|
|---|
| 568 | if(n == 0) return;
|
|---|
| 569 |
|
|---|
| 570 | copy_to_user(to, from, n);
|
|---|
| 571 | }
|
|---|
| 572 | //******************************************************************************
|
|---|
| 573 | //******************************************************************************
|
|---|
| 574 | unsigned long copy_from_user(void *to, const void *from, unsigned long n)
|
|---|
| 575 | {
|
|---|
| 576 | if(to == NULL || from == NULL) {
|
|---|
| 577 | DebugInt3();
|
|---|
| 578 | return 0;
|
|---|
| 579 | }
|
|---|
| 580 | if(n == 0) return 0;
|
|---|
| 581 |
|
|---|
| 582 | memcpy(to, from, n);
|
|---|
| 583 | return 0;
|
|---|
| 584 | }
|
|---|
| 585 | //******************************************************************************
|
|---|
| 586 | //******************************************************************************
|
|---|
| 587 | int __get_user(int size, void *dest, void *src)
|
|---|
| 588 | {
|
|---|
| 589 | if(size == 0) return 0;
|
|---|
| 590 |
|
|---|
| 591 | if(dest == NULL || src == NULL) {
|
|---|
| 592 | DebugInt3();
|
|---|
| 593 | return 0;
|
|---|
| 594 | }
|
|---|
| 595 | memcpy(dest, src, size);
|
|---|
| 596 | return 0;
|
|---|
| 597 | }
|
|---|
| 598 | //******************************************************************************
|
|---|
| 599 | //******************************************************************************
|
|---|
| 600 | int _put_user(int size, int x, void *ptr)
|
|---|
| 601 | {
|
|---|
| 602 | if(ptr == NULL || size == 0) {
|
|---|
| 603 | DebugInt3();
|
|---|
| 604 | return 0;
|
|---|
| 605 | }
|
|---|
| 606 |
|
|---|
| 607 | *(int *)ptr = x;
|
|---|
| 608 | return 0;
|
|---|
| 609 | }
|
|---|
| 610 |
|
|---|
| 611 | //******************************************************************************
|
|---|
| 612 | #ifdef DEBUGHEAP
|
|---|
| 613 | void *__kmalloc(int size, int flags, const char *filename, int lineno)
|
|---|
| 614 | #else
|
|---|
| 615 | void *__kmalloc(int size, int flags)
|
|---|
| 616 | #endif
|
|---|
| 617 | {
|
|---|
| 618 | LINEAR addr;
|
|---|
| 619 |
|
|---|
| 620 | if(size == 0) {
|
|---|
| 621 | DebugInt3();
|
|---|
| 622 | return NULL;
|
|---|
| 623 | }
|
|---|
| 624 | if(flags & GFP_DMA) {
|
|---|
| 625 | DebugInt3();
|
|---|
| 626 | }
|
|---|
| 627 | if(size >= 4096) {
|
|---|
| 628 | return vmalloc(size);
|
|---|
| 629 | }
|
|---|
| 630 | #ifdef DEBUGHEAP
|
|---|
| 631 | addr = (LINEAR)malloc(size, filename, lineno);
|
|---|
| 632 | #else
|
|---|
| 633 | addr = (LINEAR)malloc(size);
|
|---|
| 634 | #endif
|
|---|
| 635 | if(addr == NULL) {
|
|---|
| 636 | DebugInt3();
|
|---|
| 637 | return 0;
|
|---|
| 638 | }
|
|---|
| 639 | //dprintf(("kmalloc %d returned %x", size, addr));
|
|---|
| 640 | return addr;
|
|---|
| 641 | }
|
|---|
| 642 |
|
|---|
| 643 | //******************************************************************************
|
|---|
| 644 | #ifdef DEBUGHEAP
|
|---|
| 645 | void __kfree(const void *ptr, const char *filename, int lineno)
|
|---|
| 646 | #else
|
|---|
| 647 | void __kfree(const void *ptr)
|
|---|
| 648 | #endif
|
|---|
| 649 | {
|
|---|
| 650 | ULONG addr;
|
|---|
| 651 |
|
|---|
| 652 | addr = (ULONG)ptr;
|
|---|
| 653 | if(addr == 0)
|
|---|
| 654 | {
|
|---|
| 655 | DebugInt3();
|
|---|
| 656 | return;
|
|---|
| 657 | }
|
|---|
| 658 | //dprintf(("kfree %x", addr));
|
|---|
| 659 | if(IsHeapAddr(addr)) {
|
|---|
| 660 | #ifdef DEBUGHEAP
|
|---|
| 661 | free((void *)addr, filename, lineno);
|
|---|
| 662 | #else
|
|---|
| 663 | free((void *)addr);
|
|---|
| 664 | #endif
|
|---|
| 665 | }
|
|---|
| 666 | else vfree((PVOID)addr);
|
|---|
| 667 | }
|
|---|
| 668 |
|
|---|
| 669 | //******************************************************************************
|
|---|
| 670 | void *kzalloc(size_t size, unsigned int flags)
|
|---|
| 671 | {
|
|---|
| 672 | void *ret;
|
|---|
| 673 | ret = _kmalloc(size, flags);
|
|---|
| 674 | if (ret)
|
|---|
| 675 | memset(ret, 0, size);
|
|---|
| 676 | return ret;
|
|---|
| 677 | }
|
|---|
| 678 | //******************************************************************************
|
|---|
| 679 | //******************************************************************************
|
|---|
| 680 | void *kcalloc(size_t n, size_t size, unsigned int flags)
|
|---|
| 681 | {
|
|---|
| 682 | if (n != 0 && size > INT_MAX / n)
|
|---|
| 683 | return NULL;
|
|---|
| 684 | return kzalloc(n * size, flags);
|
|---|
| 685 | }
|
|---|
| 686 | //******************************************************************************
|
|---|
| 687 | //******************************************************************************
|
|---|
| 688 |
|
|---|
| 689 | size_t ksize(const void *block)
|
|---|
| 690 | {
|
|---|
| 691 | size_t size;
|
|---|
| 692 |
|
|---|
| 693 | if (!block)
|
|---|
| 694 | size = 0; // Bad coder
|
|---|
| 695 |
|
|---|
| 696 | else if (block == ZERO_SIZE_PTR)
|
|---|
| 697 | size = 0; // Bad coder
|
|---|
| 698 |
|
|---|
| 699 | else if(IsHeapAddr((ULONG)block))
|
|---|
| 700 | size = _msize((void _near *)block);
|
|---|
| 701 |
|
|---|
| 702 | else if (!GetBaseAddressNoFree((ULONG)block, (ULONG NEAR *)&size))
|
|---|
| 703 | size = 0; // Something wrong
|
|---|
| 704 |
|
|---|
| 705 | return size;
|
|---|
| 706 | }
|
|---|
| 707 | //******************************************************************************
|
|---|
| 708 | //******************************************************************************
|
|---|
| 709 | static inline void *__do_krealloc(const void *p, size_t new_size,
|
|---|
| 710 | gfp_t flags)
|
|---|
| 711 | {
|
|---|
| 712 | void *ret;
|
|---|
| 713 | size_t ks = 0;
|
|---|
| 714 |
|
|---|
| 715 | if (p)
|
|---|
| 716 | ks = ksize(p);
|
|---|
| 717 |
|
|---|
| 718 | if (ks >= new_size)
|
|---|
| 719 | return (void *)p;
|
|---|
| 720 |
|
|---|
| 721 | ret = __kmalloc(new_size, flags);
|
|---|
| 722 | if (ret && p)
|
|---|
| 723 | memcpy(ret, p, ks);
|
|---|
| 724 |
|
|---|
| 725 | return ret;
|
|---|
| 726 | }
|
|---|
| 727 | //******************************************************************************
|
|---|
| 728 | //******************************************************************************
|
|---|
| 729 | /**
|
|---|
| 730 | * krealloc - reallocate memory. The contents will remain unchanged.
|
|---|
| 731 | * @p: object to reallocate memory for.
|
|---|
| 732 | * @new_size: how many bytes of memory are required.
|
|---|
| 733 | * @flags: the type of memory to allocate.
|
|---|
| 734 | *
|
|---|
| 735 | * The contents of the object pointed to are preserved up to the
|
|---|
| 736 | * lesser of the new and old sizes. If @p is %NULL, krealloc()
|
|---|
| 737 | * behaves exactly like kmalloc(). If @new_size is 0 and @p is not a
|
|---|
| 738 | * %NULL pointer, the object pointed to is freed.
|
|---|
| 739 | */
|
|---|
| 740 | void *krealloc(const void *p, size_t new_size, gfp_t flags)
|
|---|
| 741 | {
|
|---|
| 742 | void *ret;
|
|---|
| 743 |
|
|---|
| 744 | if (!new_size) {
|
|---|
| 745 | kfree(p);
|
|---|
| 746 | return ZERO_SIZE_PTR;
|
|---|
| 747 | }
|
|---|
| 748 |
|
|---|
| 749 | ret = __do_krealloc(p, new_size, flags);
|
|---|
| 750 | if (ret && p != ret)
|
|---|
| 751 | kfree(p);
|
|---|
| 752 |
|
|---|
| 753 | return ret;
|
|---|
| 754 | }
|
|---|
| 755 | //******************************************************************************
|
|---|
| 756 | //******************************************************************************
|
|---|
| 757 | /**
|
|---|
| 758 | * vzalloc - allocate virtually contiguous memory with zero fill
|
|---|
| 759 | * @size: allocation size
|
|---|
| 760 | * Allocate enough pages to cover @size from the page level
|
|---|
| 761 | * allocator and map them into contiguous kernel virtual space.
|
|---|
| 762 | * The memory allocated is set to zero.
|
|---|
| 763 | *
|
|---|
| 764 | * For tight control over page level allocator and protection flags
|
|---|
| 765 | * use __vmalloc() instead.
|
|---|
| 766 | */
|
|---|
| 767 | void *vzalloc(unsigned long size)
|
|---|
| 768 | {
|
|---|
| 769 | void *buf;
|
|---|
| 770 | buf = vmalloc(size);
|
|---|
| 771 | if (buf)
|
|---|
| 772 | memset(buf, 0, size);
|
|---|
| 773 | return buf;
|
|---|
| 774 | }
|
|---|
| 775 | //******************************************************************************
|
|---|
| 776 | //******************************************************************************
|
|---|
| 777 | /**
|
|---|
| 778 | * kvmalloc_node - attempt to allocate physically contiguous memory, but upon
|
|---|
| 779 | * failure, fall back to non-contiguous (vmalloc) allocation.
|
|---|
| 780 | * @size: size of the request.
|
|---|
| 781 | * @flags: gfp mask for the allocation - must be compatible (superset) with GFP_KERNEL.
|
|---|
| 782 | * @node: numa node to allocate from
|
|---|
| 783 | *
|
|---|
| 784 | * Uses kmalloc to get the memory but if the allocation fails then falls back
|
|---|
| 785 | * to the vmalloc allocator. Use kvfree for freeing the memory.
|
|---|
| 786 | *
|
|---|
| 787 | * Reclaim modifiers - __GFP_NORETRY and __GFP_NOFAIL are not supported.
|
|---|
| 788 | * __GFP_RETRY_MAYFAIL is supported, and it should be used only if kmalloc is
|
|---|
| 789 | * preferable to the vmalloc fallback, due to visible performance drawbacks.
|
|---|
| 790 | *
|
|---|
| 791 | * Please note that any use of gfp flags outside of GFP_KERNEL is careful to not
|
|---|
| 792 | * fall back to vmalloc.
|
|---|
| 793 | *
|
|---|
| 794 | * Return: pointer to the allocated memory of %NULL in case of failure
|
|---|
| 795 | */
|
|---|
| 796 | void *kvmalloc_node(size_t size, gfp_t flags, int node)
|
|---|
| 797 | {
|
|---|
| 798 | gfp_t kmalloc_flags = flags;
|
|---|
| 799 | void *ret;
|
|---|
| 800 |
|
|---|
| 801 | /*
|
|---|
| 802 | * vmalloc uses GFP_KERNEL for some internal allocations (e.g page tables)
|
|---|
| 803 | * so the given set of flags has to be compatible.
|
|---|
| 804 | */
|
|---|
| 805 | if ((flags & GFP_KERNEL) != GFP_KERNEL)
|
|---|
| 806 | return kmalloc_node(size, flags, node);
|
|---|
| 807 |
|
|---|
| 808 | /*
|
|---|
| 809 | * We want to attempt a large physically contiguous block first because
|
|---|
| 810 | * it is less likely to fragment multiple larger blocks and therefore
|
|---|
| 811 | * contribute to a long term fragmentation less than vmalloc fallback.
|
|---|
| 812 | * However make sure that larger requests are not too disruptive - no
|
|---|
| 813 | * OOM killer and no allocation failure warnings as we have a fallback.
|
|---|
| 814 | */
|
|---|
| 815 | if (size > PAGE_SIZE) {
|
|---|
| 816 | kmalloc_flags |= __GFP_NOWARN;
|
|---|
| 817 |
|
|---|
| 818 | if (!(kmalloc_flags & __GFP_RETRY_MAYFAIL))
|
|---|
| 819 | kmalloc_flags |= __GFP_NORETRY;
|
|---|
| 820 | }
|
|---|
| 821 |
|
|---|
| 822 | ret = kmalloc_node(size, kmalloc_flags, node);
|
|---|
| 823 |
|
|---|
| 824 | /*
|
|---|
| 825 | * It doesn't really make sense to fallback to vmalloc for sub page
|
|---|
| 826 | * requests
|
|---|
| 827 | */
|
|---|
| 828 | if (ret || size <= PAGE_SIZE)
|
|---|
| 829 | return ret;
|
|---|
| 830 |
|
|---|
| 831 | return __vmalloc_node(size, 1, flags, node,
|
|---|
| 832 | __builtin_return_address(0));
|
|---|
| 833 | }
|
|---|
| 834 | //******************************************************************************
|
|---|
| 835 | //******************************************************************************
|
|---|