mirror of
https://github.com/adulau/aha.git
synced 2024-12-29 04:06:22 +00:00
94bf5ceac0
Add an example of how to use the MAP_HUGETLB flag to the vm documentation directory and a reference to the example in hugetlbpage.txt. Signed-off-by: Eric B Munson <ebmunson@us.ibm.com> Acked-by: David Rientjes <rientjes@google.com> Cc: Mel Gorman <mel@csn.ul.ie> Cc: Adam Litke <agl@us.ibm.com> Cc: David Gibson <david@gibson.dropbear.id.au> Cc: Lee Schermerhorn <lee.schermerhorn@hp.com> Cc: Nick Piggin <nickpiggin@yahoo.com.au> Cc: Hugh Dickins <hugh.dickins@tiscali.co.uk> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
382 lines
14 KiB
Text
382 lines
14 KiB
Text
|
|
The intent of this file is to give a brief summary of hugetlbpage support in
|
|
the Linux kernel. This support is built on top of multiple page size support
|
|
that is provided by most modern architectures. For example, i386
|
|
architecture supports 4K and 4M (2M in PAE mode) page sizes, ia64
|
|
architecture supports multiple page sizes 4K, 8K, 64K, 256K, 1M, 4M, 16M,
|
|
256M and ppc64 supports 4K and 16M. A TLB is a cache of virtual-to-physical
|
|
translations. Typically this is a very scarce resource on processor.
|
|
Operating systems try to make best use of limited number of TLB resources.
|
|
This optimization is more critical now as bigger and bigger physical memories
|
|
(several GBs) are more readily available.
|
|
|
|
Users can use the huge page support in Linux kernel by either using the mmap
|
|
system call or standard SYSv shared memory system calls (shmget, shmat).
|
|
|
|
First the Linux kernel needs to be built with the CONFIG_HUGETLBFS
|
|
(present under "File systems") and CONFIG_HUGETLB_PAGE (selected
|
|
automatically when CONFIG_HUGETLBFS is selected) configuration
|
|
options.
|
|
|
|
The kernel built with huge page support should show the number of configured
|
|
huge pages in the system by running the "cat /proc/meminfo" command.
|
|
|
|
/proc/meminfo also provides information about the total number of hugetlb
|
|
pages configured in the kernel. It also displays information about the
|
|
number of free hugetlb pages at any time. It also displays information about
|
|
the configured huge page size - this is needed for generating the proper
|
|
alignment and size of the arguments to the above system calls.
|
|
|
|
The output of "cat /proc/meminfo" will have lines like:
|
|
|
|
.....
|
|
HugePages_Total: vvv
|
|
HugePages_Free: www
|
|
HugePages_Rsvd: xxx
|
|
HugePages_Surp: yyy
|
|
Hugepagesize: zzz kB
|
|
|
|
where:
|
|
HugePages_Total is the size of the pool of huge pages.
|
|
HugePages_Free is the number of huge pages in the pool that are not yet
|
|
allocated.
|
|
HugePages_Rsvd is short for "reserved," and is the number of huge pages for
|
|
which a commitment to allocate from the pool has been made,
|
|
but no allocation has yet been made. Reserved huge pages
|
|
guarantee that an application will be able to allocate a
|
|
huge page from the pool of huge pages at fault time.
|
|
HugePages_Surp is short for "surplus," and is the number of huge pages in
|
|
the pool above the value in /proc/sys/vm/nr_hugepages. The
|
|
maximum number of surplus huge pages is controlled by
|
|
/proc/sys/vm/nr_overcommit_hugepages.
|
|
|
|
/proc/filesystems should also show a filesystem of type "hugetlbfs" configured
|
|
in the kernel.
|
|
|
|
/proc/sys/vm/nr_hugepages indicates the current number of configured hugetlb
|
|
pages in the kernel. Super user can dynamically request more (or free some
|
|
pre-configured) huge pages.
|
|
The allocation (or deallocation) of hugetlb pages is possible only if there are
|
|
enough physically contiguous free pages in system (freeing of huge pages is
|
|
possible only if there are enough hugetlb pages free that can be transferred
|
|
back to regular memory pool).
|
|
|
|
Pages that are used as hugetlb pages are reserved inside the kernel and cannot
|
|
be used for other purposes.
|
|
|
|
Once the kernel with Hugetlb page support is built and running, a user can
|
|
use either the mmap system call or shared memory system calls to start using
|
|
the huge pages. It is required that the system administrator preallocate
|
|
enough memory for huge page purposes.
|
|
|
|
The administrator can preallocate huge pages on the kernel boot command line by
|
|
specifying the "hugepages=N" parameter, where 'N' = the number of huge pages
|
|
requested. This is the most reliable method for preallocating huge pages as
|
|
memory has not yet become fragmented.
|
|
|
|
Some platforms support multiple huge page sizes. To preallocate huge pages
|
|
of a specific size, one must preceed the huge pages boot command parameters
|
|
with a huge page size selection parameter "hugepagesz=<size>". <size> must
|
|
be specified in bytes with optional scale suffix [kKmMgG]. The default huge
|
|
page size may be selected with the "default_hugepagesz=<size>" boot parameter.
|
|
|
|
/proc/sys/vm/nr_hugepages indicates the current number of configured [default
|
|
size] hugetlb pages in the kernel. Super user can dynamically request more
|
|
(or free some pre-configured) huge pages.
|
|
|
|
Use the following command to dynamically allocate/deallocate default sized
|
|
huge pages:
|
|
|
|
echo 20 > /proc/sys/vm/nr_hugepages
|
|
|
|
This command will try to configure 20 default sized huge pages in the system.
|
|
On a NUMA platform, the kernel will attempt to distribute the huge page pool
|
|
over the all on-line nodes. These huge pages, allocated when nr_hugepages
|
|
is increased, are called "persistent huge pages".
|
|
|
|
The success or failure of huge page allocation depends on the amount of
|
|
physically contiguous memory that is preset in system at the time of the
|
|
allocation attempt. If the kernel is unable to allocate huge pages from
|
|
some nodes in a NUMA system, it will attempt to make up the difference by
|
|
allocating extra pages on other nodes with sufficient available contiguous
|
|
memory, if any.
|
|
|
|
System administrators may want to put this command in one of the local rc init
|
|
files. This will enable the kernel to request huge pages early in the boot
|
|
process when the possibility of getting physical contiguous pages is still
|
|
very high. Administrators can verify the number of huge pages actually
|
|
allocated by checking the sysctl or meminfo. To check the per node
|
|
distribution of huge pages in a NUMA system, use:
|
|
|
|
cat /sys/devices/system/node/node*/meminfo | fgrep Huge
|
|
|
|
/proc/sys/vm/nr_overcommit_hugepages specifies how large the pool of
|
|
huge pages can grow, if more huge pages than /proc/sys/vm/nr_hugepages are
|
|
requested by applications. Writing any non-zero value into this file
|
|
indicates that the hugetlb subsystem is allowed to try to obtain "surplus"
|
|
huge pages from the buddy allocator, when the normal pool is exhausted. As
|
|
these surplus huge pages go out of use, they are freed back to the buddy
|
|
allocator.
|
|
|
|
When increasing the huge page pool size via nr_hugepages, any surplus
|
|
pages will first be promoted to persistent huge pages. Then, additional
|
|
huge pages will be allocated, if necessary and if possible, to fulfill
|
|
the new huge page pool size.
|
|
|
|
The administrator may shrink the pool of preallocated huge pages for
|
|
the default huge page size by setting the nr_hugepages sysctl to a
|
|
smaller value. The kernel will attempt to balance the freeing of huge pages
|
|
across all on-line nodes. Any free huge pages on the selected nodes will
|
|
be freed back to the buddy allocator.
|
|
|
|
Caveat: Shrinking the pool via nr_hugepages such that it becomes less
|
|
than the number of huge pages in use will convert the balance to surplus
|
|
huge pages even if it would exceed the overcommit value. As long as
|
|
this condition holds, however, no more surplus huge pages will be
|
|
allowed on the system until one of the two sysctls are increased
|
|
sufficiently, or the surplus huge pages go out of use and are freed.
|
|
|
|
With support for multiple huge page pools at run-time available, much of
|
|
the huge page userspace interface has been duplicated in sysfs. The above
|
|
information applies to the default huge page size which will be
|
|
controlled by the /proc interfaces for backwards compatibility. The root
|
|
huge page control directory in sysfs is:
|
|
|
|
/sys/kernel/mm/hugepages
|
|
|
|
For each huge page size supported by the running kernel, a subdirectory
|
|
will exist, of the form
|
|
|
|
hugepages-${size}kB
|
|
|
|
Inside each of these directories, the same set of files will exist:
|
|
|
|
nr_hugepages
|
|
nr_overcommit_hugepages
|
|
free_hugepages
|
|
resv_hugepages
|
|
surplus_hugepages
|
|
|
|
which function as described above for the default huge page-sized case.
|
|
|
|
If the user applications are going to request huge pages using mmap system
|
|
call, then it is required that system administrator mount a file system of
|
|
type hugetlbfs:
|
|
|
|
mount -t hugetlbfs \
|
|
-o uid=<value>,gid=<value>,mode=<value>,size=<value>,nr_inodes=<value> \
|
|
none /mnt/huge
|
|
|
|
This command mounts a (pseudo) filesystem of type hugetlbfs on the directory
|
|
/mnt/huge. Any files created on /mnt/huge uses huge pages. The uid and gid
|
|
options sets the owner and group of the root of the file system. By default
|
|
the uid and gid of the current process are taken. The mode option sets the
|
|
mode of root of file system to value & 0777. This value is given in octal.
|
|
By default the value 0755 is picked. The size option sets the maximum value of
|
|
memory (huge pages) allowed for that filesystem (/mnt/huge). The size is
|
|
rounded down to HPAGE_SIZE. The option nr_inodes sets the maximum number of
|
|
inodes that /mnt/huge can use. If the size or nr_inodes option is not
|
|
provided on command line then no limits are set. For size and nr_inodes
|
|
options, you can use [G|g]/[M|m]/[K|k] to represent giga/mega/kilo. For
|
|
example, size=2K has the same meaning as size=2048.
|
|
|
|
While read system calls are supported on files that reside on hugetlb
|
|
file systems, write system calls are not.
|
|
|
|
Regular chown, chgrp, and chmod commands (with right permissions) could be
|
|
used to change the file attributes on hugetlbfs.
|
|
|
|
Also, it is important to note that no such mount command is required if the
|
|
applications are going to use only shmat/shmget system calls or mmap with
|
|
MAP_HUGETLB. Users who wish to use hugetlb page via shared memory segment
|
|
should be a member of a supplementary group and system admin needs to
|
|
configure that gid into /proc/sys/vm/hugetlb_shm_group. It is possible for
|
|
same or different applications to use any combination of mmaps and shm*
|
|
calls, though the mount of filesystem will be required for using mmap calls
|
|
without MAP_HUGETLB. For an example of how to use mmap with MAP_HUGETLB see
|
|
map_hugetlb.c.
|
|
|
|
*******************************************************************
|
|
|
|
/*
|
|
* Example of using huge page memory in a user application using Sys V shared
|
|
* memory system calls. In this example the app is requesting 256MB of
|
|
* memory that is backed by huge pages. The application uses the flag
|
|
* SHM_HUGETLB in the shmget system call to inform the kernel that it is
|
|
* requesting huge pages.
|
|
*
|
|
* For the ia64 architecture, the Linux kernel reserves Region number 4 for
|
|
* huge pages. That means the addresses starting with 0x800000... will need
|
|
* to be specified. Specifying a fixed address is not required on ppc64,
|
|
* i386 or x86_64.
|
|
*
|
|
* Note: The default shared memory limit is quite low on many kernels,
|
|
* you may need to increase it via:
|
|
*
|
|
* echo 268435456 > /proc/sys/kernel/shmmax
|
|
*
|
|
* This will increase the maximum size per shared memory segment to 256MB.
|
|
* The other limit that you will hit eventually is shmall which is the
|
|
* total amount of shared memory in pages. To set it to 16GB on a system
|
|
* with a 4kB pagesize do:
|
|
*
|
|
* echo 4194304 > /proc/sys/kernel/shmall
|
|
*/
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <sys/types.h>
|
|
#include <sys/ipc.h>
|
|
#include <sys/shm.h>
|
|
#include <sys/mman.h>
|
|
|
|
#ifndef SHM_HUGETLB
|
|
#define SHM_HUGETLB 04000
|
|
#endif
|
|
|
|
#define LENGTH (256UL*1024*1024)
|
|
|
|
#define dprintf(x) printf(x)
|
|
|
|
/* Only ia64 requires this */
|
|
#ifdef __ia64__
|
|
#define ADDR (void *)(0x8000000000000000UL)
|
|
#define SHMAT_FLAGS (SHM_RND)
|
|
#else
|
|
#define ADDR (void *)(0x0UL)
|
|
#define SHMAT_FLAGS (0)
|
|
#endif
|
|
|
|
int main(void)
|
|
{
|
|
int shmid;
|
|
unsigned long i;
|
|
char *shmaddr;
|
|
|
|
if ((shmid = shmget(2, LENGTH,
|
|
SHM_HUGETLB | IPC_CREAT | SHM_R | SHM_W)) < 0) {
|
|
perror("shmget");
|
|
exit(1);
|
|
}
|
|
printf("shmid: 0x%x\n", shmid);
|
|
|
|
shmaddr = shmat(shmid, ADDR, SHMAT_FLAGS);
|
|
if (shmaddr == (char *)-1) {
|
|
perror("Shared memory attach failure");
|
|
shmctl(shmid, IPC_RMID, NULL);
|
|
exit(2);
|
|
}
|
|
printf("shmaddr: %p\n", shmaddr);
|
|
|
|
dprintf("Starting the writes:\n");
|
|
for (i = 0; i < LENGTH; i++) {
|
|
shmaddr[i] = (char)(i);
|
|
if (!(i % (1024 * 1024)))
|
|
dprintf(".");
|
|
}
|
|
dprintf("\n");
|
|
|
|
dprintf("Starting the Check...");
|
|
for (i = 0; i < LENGTH; i++)
|
|
if (shmaddr[i] != (char)i)
|
|
printf("\nIndex %lu mismatched\n", i);
|
|
dprintf("Done.\n");
|
|
|
|
if (shmdt((const void *)shmaddr) != 0) {
|
|
perror("Detach failure");
|
|
shmctl(shmid, IPC_RMID, NULL);
|
|
exit(3);
|
|
}
|
|
|
|
shmctl(shmid, IPC_RMID, NULL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
*******************************************************************
|
|
|
|
/*
|
|
* Example of using huge page memory in a user application using the mmap
|
|
* system call. Before running this application, make sure that the
|
|
* administrator has mounted the hugetlbfs filesystem (on some directory
|
|
* like /mnt) using the command mount -t hugetlbfs nodev /mnt. In this
|
|
* example, the app is requesting memory of size 256MB that is backed by
|
|
* huge pages.
|
|
*
|
|
* For ia64 architecture, Linux kernel reserves Region number 4 for huge pages.
|
|
* That means the addresses starting with 0x800000... will need to be
|
|
* specified. Specifying a fixed address is not required on ppc64, i386
|
|
* or x86_64.
|
|
*/
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <unistd.h>
|
|
#include <sys/mman.h>
|
|
#include <fcntl.h>
|
|
|
|
#define FILE_NAME "/mnt/hugepagefile"
|
|
#define LENGTH (256UL*1024*1024)
|
|
#define PROTECTION (PROT_READ | PROT_WRITE)
|
|
|
|
/* Only ia64 requires this */
|
|
#ifdef __ia64__
|
|
#define ADDR (void *)(0x8000000000000000UL)
|
|
#define FLAGS (MAP_SHARED | MAP_FIXED)
|
|
#else
|
|
#define ADDR (void *)(0x0UL)
|
|
#define FLAGS (MAP_SHARED)
|
|
#endif
|
|
|
|
void check_bytes(char *addr)
|
|
{
|
|
printf("First hex is %x\n", *((unsigned int *)addr));
|
|
}
|
|
|
|
void write_bytes(char *addr)
|
|
{
|
|
unsigned long i;
|
|
|
|
for (i = 0; i < LENGTH; i++)
|
|
*(addr + i) = (char)i;
|
|
}
|
|
|
|
void read_bytes(char *addr)
|
|
{
|
|
unsigned long i;
|
|
|
|
check_bytes(addr);
|
|
for (i = 0; i < LENGTH; i++)
|
|
if (*(addr + i) != (char)i) {
|
|
printf("Mismatch at %lu\n", i);
|
|
break;
|
|
}
|
|
}
|
|
|
|
int main(void)
|
|
{
|
|
void *addr;
|
|
int fd;
|
|
|
|
fd = open(FILE_NAME, O_CREAT | O_RDWR, 0755);
|
|
if (fd < 0) {
|
|
perror("Open failed");
|
|
exit(1);
|
|
}
|
|
|
|
addr = mmap(ADDR, LENGTH, PROTECTION, FLAGS, fd, 0);
|
|
if (addr == MAP_FAILED) {
|
|
perror("mmap");
|
|
unlink(FILE_NAME);
|
|
exit(1);
|
|
}
|
|
|
|
printf("Returned address is %p\n", addr);
|
|
check_bytes(addr);
|
|
write_bytes(addr);
|
|
read_bytes(addr);
|
|
|
|
munmap(addr, LENGTH);
|
|
close(fd);
|
|
unlink(FILE_NAME);
|
|
|
|
return 0;
|
|
}
|