NUMA结构的alloc_pages
    1. ==================== mm/numa.c 43 43 ====================
    1. 43 #ifdef CONFIG_DISCONTIGMEM
    1. ==================== mm/numa.c 91 128 ====================
    1. 91 /*
    1. 92 * This can be refined. Currently, tries to do round robin, instead
    1. 93 * should do concentratic circle search, starting from current node.
    1. 94 */
    1. //分配策略, 所需物理块的大小,2的order次方
    1. 95 struct page * alloc_pages(int gfp_mask, unsigned long order)
    1. 96 {
    1. 97 struct page *ret = 0;
    1. 98 pg_data_t *start, *temp;
    1. 99 #ifndef CONFIG_NUMA
    1. 100 unsigned long flags;
    1. 101 static pg_data_t *next = 0;
    1. 102 #endif
    1. 103
    1. 104 if (order >= MAX_ORDER)
    1. 105 return NULL;
    1. 106 #ifdef CONFIG_NUMA//NUMA结构
    1. 107 temp = NODE_DATA(numa_node_id());//可以通过宏操作找到cpu的节数据结构队列
    1. 108 #else
    1. 109 spin_lock_irqsave(&node_lock, flags);
    1. 110 if (!next) next = pgdat_list;
    1. 111 temp = next;
    1. 112 next = next->node_next;
    1. 113 spin_unlock_irqrestore(&node_lock, flags);
    1. 114 #endif
    1. /*
    1. 函数主要操作2个循环,一个从temp到队列末尾,一个从队头到temp,扫描所有节,直到某节点内存分配成功
    1. */
    1. 115 start = temp;
    1. 116 while (temp) {
    1. 117 if ((ret = alloc_pages_pgdat(temp, gfp_mask, order)))//接下来解析此函数
    1. 118 return(ret);
    1. 119 temp = temp->node_next;
    1. 120 }
    1. 121 temp = pgdat_list;
    1. 122 while (temp != start) {
    1. 123 if ((ret = alloc_pages_pgdat(temp, gfp_mask, order)))
    1. 124 return(ret);
    1. 125 temp = temp->node_next;
    1. 126 }
    1. 127 return(0);
    1. 128 }

alloc_pages_pgdat试图分配所需页面,是__alloc_pages的封装
    1. 85 static struct page * alloc_pages_pgdat(pg_data_t *pgdat, int gfp_mask,
    1. 86 unsigned long order)
    1. 87 { //node_zonelist决定分配策略数组
    1. 88     return __alloc_pages(pgdat->node_zonelists + gfp_mask, order);
    1. 89 }
与UMA的alloc_pages()相比较,UMA只有一个节点,contig_page_data.UMA与NUMA共同使用__alloc_pages
    1. ==================== include/linux/mm.h 343 352 ====================
    1. 343 #ifndef CONFIG_DISCONTIGMEM//只有这个无定义,才使用uma的__alloc_pages
    1. 344 static inline struct page * alloc_pages(int gfp_mask, unsigned long order)
    1. 83
    1. 345 {
    1. 346 /*
    1. 347 * Gets optimized away by the compiler.
    1. 348 */
    1. 349 if (order >= MAX_ORDER)
    1. 350 return NULL;
    1. 351 return __alloc_pages(contig_page_data.node_zonelists+(gfp_mask), order);
    1. 352 }
查看__alloc_page,
  1. 如果只分配一个页面,而且要等待完成分配,又不适用于管理的目的
  2. 把direct_reclaim设置为1,表示可以从相应的管理区的不活跃干净页面缓冲队列中回收
  3. 84发现空闲页面短缺,唤醒以下2个进程,试图腾出一些页面出来
    1. [alloc_pages()>__alloc_pages()]
    1. 270 /*
    1. 271 * This is the 'heart' of the zoned buddy allocator:
    1. 272 */
    1. 273 struct page * __alloc_pages(zonelist_t *zonelist, unsigned long order)
    1. 274 {
    1. 275 zone_t **zone;
    1. 276 int direct_reclaim = 0;
    1. 277 unsigned int gfp_mask = zonelist->gfp_mask;//获取具体的分配策略
    1. 278 struct page * page;
    1. 279
    1. 280 /*
    1. 281 * Allocations put pressure on the VM subsystem.
    1. 282 */
    1. 283 memory_pressure++;//表示内存管理所承受的压力,分配++,归还--
    1. 284
    1. 285 /*
    1. 286 * (If anyone calls gfp from interrupts nonatomically then it
    1. 287 * will sooner or later tripped up by a schedule().)
    1. 288 *
    1. 289 * We are falling back to lower-level zones if allocation
    1. 290 * in a higher zone fails.
    1. 291 */
    1. 292
    1. 293 /*
    1. 294 如果只分配一个页面,而且要等待完成分配,又不适用于管理的目的
    1. 把direct_reclaim设置为1,表示可以从相应的管理区的不活跃干净页面缓冲队列中回收
    1. 296 */
    1. 297 if (order == 0 && (gfp_mask & __GFP_WAIT) &&
    1. 298 !(current->flags & PF_MEMALLOC))
    1. 299 direct_reclaim = 1;
    1. 300
    1. 301 /*
    1. 302 * If we are about to get low on free pages and we also have
    1. 303 * an inactive page shortage, wake up kswapd.
    1. 84发现空闲页面短缺,唤醒以下2个进程,试图腾出一些页面出来
    1. 304 */
    1. 305 if (inactive_shortage() > inactive_target / 2 && free_shortage())
    1. 306 wakeup_kswapd(0);
    1. 307 /*
    1. 308 * If we are about to get low on free pages and cleaning
    1. 309 * the inactive_dirty pages would fix the situation,
    1. 310 * wake up bdflush.
    1. 311 */
    1. 312 else if (free_shortage() && nr_inactive_dirty_pages > free_shortage()
    1. 313 && nr_inactive_dirty_pages >= freepages.high)
    1. 314 wakeup_bdflush(0);
    1. 315
继续看__alloc_page代码
//如果管理区的空闲页面大于其最低标准,分配成功直接返回
  1. //否则有进程(内核线程kreclaimd)在等待队列睡眠,把它唤醒,用于回收一些页面,备用
    1. ==================== mm/page_alloc.c 316 340 ====================
    1. [alloc_pages()>__alloc_pages()]
    1. 316 try_again:
    1. 317 /*
    1. 318 * First, see if we have any zones with lots of free memory.
    1. 319 *
    1. 320 * We allocate free memory first because it doesn't contain
    1. 321 * any data ... DUH!
    1. 322 */
    1. 323 zone = zonelist->zones;//获取管理区指针
    1. 324 for (;;) {
    1. 325 zone_t *z = *(zone++);//管理区
    1. 326 if (!z)
    1. 327 break;
    1. 328 if (!z->size)
    1. 329 BUG();
    1. 330//如果管理区的空闲页面大于其最低标准
    1. 331 if (z->free_pages >= z->pages_low) {
    1. 332 page = rmqueue(z, order);//分配内存,接下来分析此函数
    1. 333 if (page)
    1. 334 return page;
    1. 335 } 
    1. //否则有进程(内核线程kreclaimd)在等待队列睡眠,把它唤醒,用于回收一些页面,备用
    1. else if (z->free_pages < z->pages_min &&
    1. 336 waitqueue_active(&kreclaimd_wait)) {
    1. 85
    1. 337 wake_up_interruptible(&kreclaimd_wait);
    1. 338 }
    1. 339 }
    1. 340

    1. [alloc_pages()>__alloc_pages()>rmqueue()]
    1. 172 static struct page * rmqueue(zone_t *zone, unsigned long order)
    1. 173 {
    1. 174 free_area_t * area = zone->free_area + order;//获取其数组对应的元素
    1. 175 unsigned long curr_order = order;
    1. 176 struct list_head *head, *curr;
    1. 177 unsigned long flags;
    1. 178 struct page *page;
    1. 179
    1. 180 spin_lock_irqsave(&zone->lock, flags);//相应管理区加锁
    1. 181 do {
    1. 182 head = &area->free_list;//头
    1. 183 curr = memlist_next(head);//头的下一个节点
    1. 184
    1. 185 if (curr != head) {//不等于空,说明有物理页块
    1. 186 unsigned int index;
    1. 187//从非空队列中取出第一个结构page元素
    1. 188 page = memlist_entry(curr, struct page, list);
    1. 189 if (BAD_RANGE(zone,page))
    1. 190 BUG();
    1. 191 memlist_del(curr);//删除队列中的元素
    1. 192 index = (page - mem_map) - zone->offset;//偏移
    1. 193 MARK_USED(index, curr_order, area);//将相应位图设置为1
    1. 194 zone->free_pages -= 1 << order;
    1. 195//分配成功,把大块剩余的部分分解为小块,链入相应的队列
    1. 196 page = expand(zone, page, index, order, curr_order, area);
    1. 197 spin_unlock_irqrestore(&zone->lock, flags);
    1. 198
    1. 199 set_page_count(page, 1);
    1. 200 if (BAD_RANGE(zone,page))
    1. 201 BUG();
    1. 202 DEBUG_ADD_PAGE
    1. 203 return page;
    1. 204 }
    1. 205 curr_order++;
    1. 206 area++;
    1. 86
    1. 207 } while (curr_order < MAX_ORDER);
    1. 208 spin_unlock_irqrestore(&zone->lock, flags);
    1. 209
    1. 210 return NULL;
    1. 211 }


    1. [alloc_pages()>__alloc_pages()>rmqueue()>expand()]
    1. /*
    1.     
          low表示所需块大小,high表示实际大小
    1.     */
    1. 150 static inline struct page * expand (zone_t *zone, struct page *page,
    1. 151 unsigned long index, int low, int high, free_area_t * area)
    1. 152 {
    1. 153 unsigned long size = 1 << high;
    1. 154
    1. 155 while (high > low) {
    1. 156 if (BAD_RANGE(zone,page))
    1. 157 BUG();
    1. 158 area--;
    1. 159 high--;
    1. 160 size >>= 1;//每次减少2的n次方
    1. 161 memlist_add_head(&(page)->list, &(area)->free_list);
    1. 162 MARK_USED(index, high, area);//标记位图
    1. //处理更低一档的空闲块队列
    1. 163 index += size;
    1. 164 page += size;
    1. 165 }
    1. 166 if (BAD_RANGE(zone,page))
    1. 167 BUG();
    1. 168 return page;
    1. 169 }
就这样rmqueue队列一直往上扫描,直到分配成功或者失败,如果失败,则__alloc_pages通过for循环
指向下一个管理区(按照分配策略),直到成功.
要是给定的分配策略中的所有页面管理区都失败,那就只能加大力度再试试.要么降低对页面的水位要求
要么把缓冲在管理区的不活跃干净页面也给考虑进去
    1. [alloc_pages()>__alloc_pages()]
    1. 341 /*
    1. 342 * Try to allocate a page from a zone with a HIGH
    1. 343 * amount of free + inactive_clean pages.
    1. 344 *
    1. 345 * If there is a lot of activity, inactive_target
    1. 346 * will be high and we'll have a good chance of
    1. 347 * finding a page using the HIGH limit.
    1. 348 */
    1. //先用page_high,如果不行再用page_low
    1. 349 page = __alloc_pages_limit(zonelist, order, PAGES_HIGH, direct_reclaim);
    1. 350 if (page)
    1. 351 return page;
    1. 352
    1. 353 /*
    1. 354 * Then try to allocate a page from a zone with more
    1. 355 * than zone->pages_low free + inactive_clean pages.
    1. 356 *
    1. 357 * When the working set is very large and VM activity
    1. 358 * is low, we're most likely to have our allocation
    1. 359 * succeed here.
    1. 360 */
    1. 361 page = __alloc_pages_limit(zonelist, order, PAGES_LOW, direct_reclaim);
    1. 362 if (page)
    1. 363 return page;
    1. 364


    1. [alloc_pages()>__alloc_pages()>__alloc_pages_limit()]
    1. 213 #define PAGES_MIN 0
    1. 214 #define PAGES_LOW 1
    1. 215 #define PAGES_HIGH 2
    1. 88
    1. 216
    1. 217 /*
    1. 218 * This function does the dirty work for __alloc_pages
    1. 219 * and is separated out to keep the code size smaller.
    1. 220 * (suggested by Davem at 1:30 AM, typed by Rik at 6 AM)
    1. 221 */
    1. 222 static struct page * __alloc_pages_limit(zonelist_t *zonelist,
    1. 223 unsigned long order, int limit, int direct_reclaim)
    1. 224 {
    1. 225 zone_t **zone = zonelist->zones;
    1. 226
    1. 227 for (;;) {
    1. 228 zone_t *z = *(zone++);
    1. 229 unsigned long water_mark;
    1. 230
    1. 231 if (!z)
    1. 232 break;
    1. 233 if (!z->size)
    1. 234 BUG();
    1. 235
    1. 236 /*
    1. 237 * We allocate if the number of free + inactive_clean
    1. 238 * pages is above the watermark.
    1. 239 */
    1. 240 switch (limit) {
    1. 241 default:
    1. 242 case PAGES_MIN://通过分配策略,改变水位
    1. 243 water_mark = z->pages_min;
    1. 244 break;
    1. 245 case PAGES_LOW:
    1. 246 water_mark = z->pages_low;
    1. 247 break;
    1. 248 case PAGES_HIGH:
    1. 249 water_mark = z->pages_high;
    1. 250 }
    1. 251//如果空闲页面+干净回收页面大于最低水位
    1. 252 if (z->free_pages + z->inactive_clean_pages > water_mark) {
    1. 253 struct page *page = NULL;
    1. 254 /* 如果空闲页面小于最低水位+8,那就回收. */
    1. 255 if (direct_reclaim && z->free_pages < z->pages_min + 8)
    1. 256 page = reclaim_page(z);//把inactive_clean_list队列回收页面
    1. 257 /* If that fails, fall back to rmqueue. */
    1. 258 if (!page)
    1. 259 page = rmqueue(z, order);
    1. 260 if (page)
    1. 261 return page;
    1. 262 }
    1. 263 }
    1. 264
    1. 89
    1. 265 /* Found nothing. */
    1. 266 return NULL;
    1. 267 }
如果还是不行,那就说明管理区的页面很短缺了
    1. [alloc_pages()>__alloc_pages()]
    1. 365 /*
    1. 366 * OK, none of the zones on our zonelist has lots
    1. 367 * of pages free.
    1. 368 *
    1. 369 * We wake up kswapd, in the hope that kswapd will
    1. 370 * resolve this situation before memory gets tight.
    1. 371 *
    1. 372 * We also yield the CPU, because that:
    1. 373 * - gives kswapd a chance to do something
    1. 374 * - slows down allocations, in particular the
    1. 375 * allocations from the fast allocator that's
    1. 376 * causing the problems ...
    1. 377 * - ... which minimises the impact the "bad guys"
    1. 378 * have on the rest of the system
    1. 379 * - if we don't have __GFP_IO set, kswapd may be
    1. 380 * able to free some memory we can't free ourselves
    1. 381 */
    1. 382 wakeup_kswapd(0);//唤醒内核线程,想办法换出一些页面
    1. 383 if (gfp_mask & __GFP_WAIT) {//要求必须获取页面,分配不到时等待,那就让系统再调用一次(目的为了调度kswapd线程)
    1. //以此获取一些页面
    1. 384 __set_current_state(TASK_RUNNING);
    1. 385 current->policy |= SCHED_YIELD;
    1. 386 schedule();
    1. 387 }
    1. 388
    1. 389 /*
    1. 390 * After waking up kswapd, we try to allocate a page
    1. 391 * from any zone which isn't critical yet.
    1. 392 *
    1. 393 * Kswapd should, in most situations, bring the situation
    1. 394 * back to normal in no time.
    1. 395 */
    1. /*
    1. 如果不允许等待,那就用pages_min再调用一次__alloc_pages_limit
    1. */
    1. 396 page = __alloc_pages_limit(zonelist, order, PAGES_MIN, direct_reclaim);
    1. 397 if (page)
    1. 398 return page;
    1. 399
要是再失败,那就看谁在要求分配内存页面.如果是kswaped,本身就是内存分配工作者
是要更好的分配页面,比一般进程更重要,那就PF_memalloc标志位为1,不过我们先看一般进程
即pe_memalloc标志位为0的策略.

    1. ==================== mm/page_alloc.c 400 477 ====================
    1. [alloc_pages()>__alloc_pages()]
    1. 400 /*
    1. 401 * Damn, we didn't succeed.
    1. 402 *
    1. 403 * This can be due to 2 reasons:
    1. 404 * - we're doing a higher-order allocation
    1. 405 * --> move pages to the free list until we succeed
    1. 406 * - we're /really/ tight on memory
    1. 407 * --> wait on the kswapd waitqueue until memory is freed
    1. 408 */
    1. 409 if (!(current->flags & PF_MEMALLOC)) {
    1. 410 /*
    1. 411 * Are we dealing with a higher order allocation?
    1. 412 *
    1. 413 * Move pages from the inactive_clean to the free list
    1. 414 * in the hope of creating a large, physically contiguous
    1. 415 * piece of free memory.
    1. 416 */
    1. 417 if (order > 0 && (gfp_mask & __GFP_WAIT)) {
    1. 418 zone = zonelist->zones;
    1. 419 /* First, clean some dirty pages. */
    1. 420 current->flags |= PF_MEMALLOC;
    1. 421 page_launder(gfp_mask, 1);//把脏页洗干净(页面的定期换出)
    1. 422 current->flags &= ~PF_MEMALLOC;
    1. 423 for (;;) {
    1. 424 zone_t *z = *(zone++);//通过一个for循环把干净页面等待队列的页面回收
    1. 425 if (!z)
    1. 426 break;
    1. 427 if (!z->size)
    1. 428 continue;
    1. //是否有干净页面
    1. 429 while (z->inactive_clean_pages) {
    1. 430 struct page * page;
    1. 431 /* Move one page to the free list. */
    1. 432 page = reclaim_page(z);//回收干净页面等待队列
    1. 433 if (!page)
    1. 434 break;
    1. 91
    1. 435 __free_page(page);//通过__free_page释放页面的同时,把空闲页面拼接成大的页面块
    1. 436 /* Try if the allocation succeeds. */
    1. 437 page = rmqueue(z, order);//试图再次请求成功
    1. 438 if (page)
    1. 439 return page;
    1. 440 }
    1. 441 }
    1. 442 }
    1. 443 /*
    1. 444 * When we arrive here, we are really tight on memory.
    1. 445 *
    1. 446 * We wake up kswapd and sleep until kswapd wakes us
    1. 447 * up again. After that we loop back to the start.
    1. 448 *
    1. 449 * We have to do this because something else might eat
    1. 450 * the memory kswapd frees for us and we need to be
    1. 451 * reliable. Note that we don't loop back for higher
    1. 452 * order allocations since it is possible that kswapd
    1. 453 * simply cannot free a large enough contiguous area
    1. 454 * of memory *ever*.
    1. 455 */
    1. /*
    1. 如果依旧失败,而且必须要求分配到页面,那就等待,进程睡眠
    1. */
    1. 456 if ((gfp_mask & (__GFP_WAIT|__GFP_IO)) == (__GFP_WAIT|__GFP_IO)) {
    1. 457 wakeup_kswapd(1);//唤醒kswaped,要求分配页面进程睡眠,等待kswapd完成一轮运行再唤醒需要页面的进程
    1. 458 memory_pressure++;
    1. 459 if (!order)//如果要求分配的是1个页面,跳到try_again
    1. 460 goto try_again;
    1. 461 /*
    1. 462 * If __GFP_IO isn't set, we can't wait on kswapd because
    1. 463 * kswapd just might need some IO locks /we/ are holding ...
    1. 464 *
    1. 465 * SUBTLE: The scheduling point above makes sure that
    1. 466 * kswapd does get the chance to free memory we can't
    1. 467 * free ourselves...
    1. 468 */
    1. 469 } else if (gfp_mask & __GFP_WAIT) {
    1. 470 try_to_free_pages(gfp_mask);//另外一种方案...直接调用此函数获取页面(本来就是kswaped函数调用的)
    1. 471 memory_pressure++;
    1. 472 if (!order)
    1. 473 goto try_again;
    1. 474 }
    1. 475
    1. 476 }
    1. 477
最后的办法了
 
    1. [alloc_pages()>__alloc_pages()]
    1. 478 /*
    1. 479 * Final phase: allocate anything we can!
    1. 480 *
    1. 481 * Higher order allocations, GFP_ATOMIC allocations and
    1. 482 * recursive allocations (PF_MEMALLOC) end up here.
    1. 483 *
    1. 484 * Only recursive allocations can use the very last pages
    1. 485 * in the system, otherwise it would be just too easy to
    1. 486 * deadlock the system...
    1. 487 */
    1. 488 zone = zonelist->zones;
    1. 489 for (;;) {
    1. 490 zone_t *z = *(zone++);
    1. 491 struct page * page = NULL;
    1. 492 if (!z)
    1. 493 break;
    1. 494 if (!z->size)
    1. 495 BUG();
    1. 496
    1. 497 /*
    1. 498 * SUBTLE: direct_reclaim is only possible if the task
    1. 499 * becomes PF_MEMALLOC while looping above. This will
    1. 500 * happen when the OOM killer selects this task for
    1. 501 * instant execution...
    1. 93
    1. 502 */
    1. 503 if (direct_reclaim) {
    1. 504 page = reclaim_page(z);
    1. 505 if (page)
    1. 506 return page;
    1. 507 }
    1. 508
    1. 509 /* XXX: is pages_min/4 a good amount to reserve for this? */
    1. 510 if (z->free_pages < z->pages_min / 4 &&
    1. 511 !(current->flags & PF_MEMALLOC))
    1. 512 continue;
    1. 513 page = rmqueue(z, order);
    1. 514 if (page)
    1. 515 return page;
    1. 516 }
    1. 517
    1. 518 /* No luck.. */
    1. 519 printk(KERN_ERR "__alloc_pages: %lu-order allocation failed.\n", order);
    1. 520 return NULL;
    1. 521 }
如果这都失败,那就系统一定出现了问题了
节点->页面短缺->调用线程,试图腾出页面->
开始遍历每个管理区->一旦管理区的空闲页面大于最低水位,那就调用rmqueue进行分配,否则把kcreclaimd线程唤醒,回收页面
rmqueue分析->如果失败,换一个管理区(按照分配策略),如果全部失败->降低页面的水位要求,把不活跃干净的页面考虑进来
->调用__alloc_pages_limit->如果空闲页面小于最低水位+8,那就回收干净页面队列(换出,腾出空间)->失败,唤醒内核线程,获取页面
->依旧失败,把脏页面洗干净,换出.获取页面->依旧失败,再次调用线程换取页面,依旧失败->把水位降低到1/4看能否满足分配->
依旧不能,系统出了问题
 





Linux内核情景分析的alloc_pages的更多相关文章

  1. linux内核情景分析之execve()

    用来描述用户态的cpu寄存器在内核栈中保存情况.可以获取用户空间的信息 struct pt_regs { long ebx; //可执行文件路径的指针(regs.ebx中 long ecx; //命令 ...

  2. Linux内核情景分析之消息队列

    早期的Unix通信只有管道与信号,管道的缺点: 所载送的信息是无格式的字节流,不知道分界线在哪,也没通信规范,另外缺乏控制手段,比如保温优先级,管道机制的大小只有1页,管道很容易写满而读取没有及时,发 ...

  3. Linux内核情景分析之异常访问,用户堆栈的扩展

    情景假设: 在堆内存中申请了一块内存,然后释放掉该内存,然后再去访问这块内存.也就是所说的野指针访问. 当cpu产生页面错误时,会把失败的线性地址放在cr2寄存器.线性地址缺页异常的4种情况 1.如果 ...

  4. linux内核情景分析之exit与Wait

    //第一层系统调用 asmlinkage long sys_exit(int error_code) { do_exit((error_code&0xff)<<8); } 其主体是 ...

  5. linux内核情景分析之内核中的互斥操作

    信号量机制: struct sempahore是其结构,定义如下 struct semaphore { atomic_t count;//资源数目 int sleepers;//等待进程数目 wait ...

  6. linux内核情景分析之命名管道

    管道是一种"无名","无形文件,只可以近亲进程使用,不可以再任意两个进程通信使用,所以只能实现"有名","有形"的文件来实现就可以 ...

  7. linux内核情景分析之信号实现

    信号在进程间通信是异步的,每个进程的task_struct结构有一个sig指针,指向一个signal_struct结构 定义如下 struct signal_struct { atomic_t cou ...

  8. linux内核情景分析之强制性调度

    从系统调用返回到用户空间是否调度,从ret_with_reschedule可看出,是否真正调度,取决于当前进程的pcb中的need_resched是否设置为1,那如何设置为1取决于以下几种情况: 时间 ...

  9. linux内核情景分析之匿名管道

    管道的机制由pipe()创建,由pipe()所建立的管道两端都在同一进程.所以必须在fork的配合下,才可以在具有亲缘关系的进程通信 /* * sys_pipe() is the normal C c ...

随机推荐

  1. 笔记-爬虫-selenium常用方法

    笔记-爬虫-selenium常用方法 1.      查找元素 常用的查找方法 find_element_by_name find_element_by_xpath find_element_by_l ...

  2. PHP.13-日历类实现

    日历类实现 1.输出星期 calendar.class.php <?php class Calendar{ function out(){//输出表格 echo '<table align ...

  3. JAVA API访问Hbase org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=32

    Java使用API访问Hbase报错: 我的hbase主节点是spark1   java代码访问hbase的时候写的是ip 结果运行程序报错 不能够识别主机名 修改主机名     修改主机hosts文 ...

  4. python语法re.compile模块介绍

    1. re模块是正则表达式模块,re模块中包含一个重要函数是compile(pattern [, flags]) ,该函数根据包含的正则表达式的字符串创建模式对象.可以实现更有效率的匹配. impor ...

  5. java.math.BigDecimal cannot be cast to java.lang.String解决方法

    从mysql数据库里取decimal(18,2)封装到Map<String,String>中 BigDecimal b = new BigDecimal(resultMap.get(&qu ...

  6. JMeter学习笔记(七) 导出文件接口测试

    导出文件接口,其实跟下载文件接口的测试类似,主要就是执行接口导出文件后保存到本地. 下载文件接口测试,参考文档:https://www.cnblogs.com/xiaoyu2018/p/1017830 ...

  7. java中利用正则表达式获取a标签

    // 设置新闻内容 notice.setContent(editorValue); Matcher m = Pattern.compile("<a[^>]*>([^< ...

  8. python XlsxWriter创建Excel 表格

    文档(英文) https://xlsxwriter.readthedocs.io/index.html 常用模块说明(中文) https://blog.csdn.net/sinat_35930259/ ...

  9. HDU 4031 Attack (线段树)

    成功袭击次数=所有袭击次数-成功防守次数 需要一个辅助pre来记录上一次袭击成功什么时候,对于每个查询,从上一次袭击成功开始,每隔t更新一次. 感觉这样做最坏时间复杂度是O(n^2),这里 说是O(q ...

  10. NYOJ 简单数据结构

    NYOJ 2 括号配对问题 栈的简单应用.可使用STL. #include <iostream> #include <cstdio> #include <cstring& ...