mirror of
https://github.com/autc04/Retro68.git
synced 2024-12-03 10:49:58 +00:00
1157 lines
37 KiB
C
1157 lines
37 KiB
C
/* Copyright (C) 2021 Free Software Foundation, Inc.
|
|
Contributed by Oracle.
|
|
|
|
This file is part of GNU Binutils.
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 3, or (at your option)
|
|
any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, 51 Franklin Street - Fifth Floor, Boston,
|
|
MA 02110-1301, USA. */
|
|
|
|
#include "config.h"
|
|
#include <dlfcn.h>
|
|
#include <pthread.h>
|
|
#include <errno.h>
|
|
#include <fcntl.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <unistd.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/param.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include "gp-defs.h"
|
|
#include "collector.h"
|
|
#include "gp-experiment.h"
|
|
#include "memmgr.h"
|
|
|
|
/* TprintfT(<level>,...) definitions. Adjust per module as needed */
|
|
#define DBG_LT0 0 // for high-level configuration, unexpected errors/warnings
|
|
#define DBG_LT1 1 // for configuration details, warnings
|
|
#define DBG_LT2 2
|
|
#define DBG_LT3 3
|
|
|
|
/* ------------- Data and prototypes for block management --------- */
|
|
#define IO_BLK 0 /* Concurrent requests */
|
|
#define IO_SEQ 1 /* All requests are sequential, f.e. JAVA_CLASSES */
|
|
#define IO_TXT 2 /* Sequential requests. Text strings. */
|
|
#define ST_INIT 0 /* Initial state. Not allocated */
|
|
#define ST_FREE 1 /* Available */
|
|
#define ST_BUSY 2 /* Not available */
|
|
|
|
/* IO_BLK, IO_SEQ */
|
|
#define NCHUNKS 64
|
|
|
|
/* IO_TXT */
|
|
#define NBUFS 64 /* Number of text buffers */
|
|
#define CUR_BUSY(x) ((uint32_t) ((x)>>63)) /* bit 63 */
|
|
#define CUR_INDX(x) ((uint32_t) (((x)>>57) & 0x3fULL)) /* bits 62:57 */
|
|
#define CUR_FOFF(x) ((x) & 0x01ffffffffffffffULL) /* bits 56: 0 */
|
|
#define CUR_MAKE(busy, indx, foff) ((((uint64_t)(busy))<<63) | (((uint64_t)(indx))<<57) | ((uint64_t)(foff)) )
|
|
|
|
typedef struct Buffer
|
|
{
|
|
uint8_t *vaddr;
|
|
uint32_t left; /* bytes left */
|
|
uint32_t state; /* ST_FREE or ST_BUSY */
|
|
} Buffer;
|
|
|
|
typedef struct DataHandle
|
|
{
|
|
Pckt_type kind; /* obsolete (to be removed) */
|
|
int iotype; /* IO_BLK, IO_SEQ, IO_TXT */
|
|
int active;
|
|
char fname[MAXPATHLEN]; /* data file name */
|
|
|
|
/* IO_BLK, IO_SEQ */
|
|
uint32_t nflow; /* number of data flows */
|
|
uint32_t *blkstate; /* block states, nflow*NCHUNKS array */
|
|
uint32_t *blkoff; /* block offset, nflow*NCHUNKS array */
|
|
uint32_t nchnk; /* number of active chunks, probably small for IO_BLK */
|
|
uint8_t *chunks[NCHUNKS]; /* chunks (nflow contiguous blocks in virtual memory) */
|
|
uint32_t chblk[NCHUNKS]; /* number of active blocks in a chunk */
|
|
uint32_t nblk; /* number of blocks in data file */
|
|
int exempt; /* if exempt from experiment size limit */
|
|
|
|
/* IO_TXT */
|
|
Buffer *buffers; /* array of text buffers */
|
|
uint64_t curpos; /* current buffer and file offset */
|
|
} DataHandle;
|
|
|
|
#define PROFILE_DATAHNDL_MAX 16
|
|
static DataHandle data_hndls[PROFILE_DATAHNDL_MAX];
|
|
static int initialized = 0;
|
|
static long blksz; /* Block size. Multiple of page size. Power of two to make (x%blksz)==(x&(blksz-1)) fast. */
|
|
static long log2blksz; /* log2(blksz) to make (x/blksz)==(x>>log2blksz) fast. */
|
|
static uint32_t size_limit; /* Experiment size limit */
|
|
static uint32_t cur_size; /* Current experiment size */
|
|
static void init ();
|
|
static void deleteHandle (DataHandle *hndl);
|
|
static int exp_size_ck (int nblocks, char *fname);
|
|
|
|
/* IO_BLK, IO_SEQ */
|
|
static int allocateChunk (DataHandle *hndl, unsigned ichunk);
|
|
static uint8_t *getBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk);
|
|
static int remapBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk);
|
|
static int newBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk);
|
|
static void deleteBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk);
|
|
|
|
/* IO_TXT */
|
|
static int is_not_the_log_file (char *fname);
|
|
static int mapBuffer (char *fname, Buffer *buf, off64_t foff);
|
|
static int newBuffer (DataHandle *hndl, uint64_t pos);
|
|
static void writeBuffer (Buffer *buf, int blk_off, char *src, int len);
|
|
static void deleteBuffer (Buffer *buf);
|
|
|
|
/*
|
|
* Common buffer management routines
|
|
*/
|
|
static void
|
|
init ()
|
|
{
|
|
/* set the block size */
|
|
long pgsz = CALL_UTIL (sysconf)(_SC_PAGESIZE);
|
|
blksz = pgsz;
|
|
log2blksz = 16; /* ensure a minimum size */
|
|
while ((1 << log2blksz) < blksz)
|
|
log2blksz += 1;
|
|
blksz = 1L << log2blksz; /* ensure that blksz is a power of two */
|
|
TprintfT (DBG_LT1, "iolib init: page size=%ld (0x%lx) blksz=%ld (0x%lx) log2blksz=%ld\n",
|
|
pgsz, pgsz, (long) blksz, (long) blksz, (long) log2blksz);
|
|
size_limit = 0;
|
|
cur_size = 0;
|
|
initialized = 1;
|
|
}
|
|
|
|
DataHandle *
|
|
__collector_create_handle (char *descp)
|
|
{
|
|
int exempt = 0;
|
|
char *desc = descp;
|
|
if (desc[0] == '*')
|
|
{
|
|
desc++;
|
|
exempt = 1;
|
|
}
|
|
if (!initialized)
|
|
init ();
|
|
|
|
/* set up header for file, file name, etc. */
|
|
if (__collector_exp_dir_name == NULL)
|
|
{
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">__collector_exp_dir_name==NULL</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_EXPOPEN);
|
|
return NULL;
|
|
}
|
|
char fname[MAXPATHLEN];
|
|
CALL_UTIL (strlcpy)(fname, __collector_exp_dir_name, sizeof (fname));
|
|
CALL_UTIL (strlcat)(fname, "/", sizeof (fname));
|
|
Pckt_type kind = 0;
|
|
int iotype = IO_BLK;
|
|
if (__collector_strcmp (desc, SP_HEAPTRACE_FILE) == 0)
|
|
kind = HEAP_PCKT;
|
|
else if (__collector_strcmp (desc, SP_SYNCTRACE_FILE) == 0)
|
|
kind = SYNC_PCKT;
|
|
else if (__collector_strcmp (desc, SP_IOTRACE_FILE) == 0)
|
|
kind = IOTRACE_PCKT;
|
|
else if (__collector_strcmp (desc, SP_RACETRACE_FILE) == 0)
|
|
kind = RACE_PCKT;
|
|
else if (__collector_strcmp (desc, SP_PROFILE_FILE) == 0)
|
|
kind = PROF_PCKT;
|
|
else if (__collector_strcmp (desc, SP_OMPTRACE_FILE) == 0)
|
|
kind = OMP_PCKT;
|
|
else if (__collector_strcmp (desc, SP_HWCNTR_FILE) == 0)
|
|
kind = HW_PCKT;
|
|
else if (__collector_strcmp (desc, SP_DEADLOCK_FILE) == 0)
|
|
kind = DEADLOCK_PCKT;
|
|
else if (__collector_strcmp (desc, SP_FRINFO_FILE) == 0)
|
|
CALL_UTIL (strlcat)(fname, "data.", sizeof (fname));
|
|
else if (__collector_strcmp (desc, SP_LOG_FILE) == 0)
|
|
iotype = IO_TXT;
|
|
else if (__collector_strcmp (desc, SP_MAP_FILE) == 0)
|
|
iotype = IO_TXT;
|
|
else if (__collector_strcmp (desc, SP_JCLASSES_FILE) == 0)
|
|
iotype = IO_SEQ;
|
|
else
|
|
{
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">iolib unknown file desc %s</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_EXPOPEN, desc);
|
|
return NULL;
|
|
}
|
|
|
|
CALL_UTIL (strlcat)(fname, desc, sizeof (fname));
|
|
TprintfT (DBG_LT1, "createHandle calling open on fname = `%s', desc = `%s' %s\n",
|
|
fname, desc, (exempt == 0 ? "non-exempt" : "exempt"));
|
|
|
|
/* allocate a handle -- not mt-safe */
|
|
DataHandle *hndl = NULL;
|
|
for (int i = 0; i < PROFILE_DATAHNDL_MAX; ++i)
|
|
if (data_hndls[i].active == 0)
|
|
{
|
|
hndl = &data_hndls[i];
|
|
break;
|
|
}
|
|
|
|
/* out of handles? */
|
|
if (hndl == NULL)
|
|
{
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">%s</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_NOHNDL, fname);
|
|
return NULL;
|
|
}
|
|
|
|
hndl->kind = kind;
|
|
hndl->nblk = 0;
|
|
hndl->exempt = exempt;
|
|
CALL_UTIL (strlcpy)(hndl->fname, fname, sizeof (hndl->fname));
|
|
int fd = CALL_UTIL (open)(hndl->fname,
|
|
O_RDWR | O_CREAT | O_TRUNC | O_EXCL,
|
|
S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH);
|
|
if (fd < 0)
|
|
{
|
|
TprintfT (0, "createHandle open failed -- hndl->fname = `%s', SP_LOG_FILE = `%s': %s\n",
|
|
hndl->fname, SP_LOG_FILE, CALL_UTIL (strerror)(errno));
|
|
if (is_not_the_log_file (hndl->fname) == 0)
|
|
{
|
|
char errbuf[4096];
|
|
/* If we are trying to create the handle for the log file, write to stderr, not the experiment */
|
|
CALL_UTIL (snprintf)(errbuf, sizeof (errbuf),
|
|
"create_handle: COL_ERROR_LOG_OPEN %s: %s\n", hndl->fname, CALL_UTIL (strerror)(errno));
|
|
CALL_UTIL (write)(2, errbuf, CALL_UTIL (strlen)(errbuf));
|
|
|
|
}
|
|
else
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">%s: create_handle</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEOPN, errno, hndl->fname);
|
|
return NULL;
|
|
}
|
|
CALL_UTIL (close)(fd);
|
|
|
|
hndl->iotype = iotype;
|
|
if (hndl->iotype == IO_TXT)
|
|
{
|
|
/* allocate our buffers in virtual memory */
|
|
/* later, we will remap buffers individually to the file */
|
|
uint8_t *memory = (uint8_t*) CALL_UTIL (mmap64)(0,
|
|
(size_t) (NBUFS * blksz),
|
|
PROT_READ | PROT_WRITE,
|
|
#if ARCH(SPARC)
|
|
MAP_SHARED | MAP_ANON,
|
|
#else
|
|
MAP_PRIVATE | MAP_ANON,
|
|
#endif
|
|
-1,
|
|
(off64_t) 0);
|
|
if (memory == MAP_FAILED)
|
|
{
|
|
TprintfT (0, "create_handle: can't mmap MAP_ANON (for %s): %s\n", hndl->fname, CALL_UTIL (strerror)(errno));
|
|
/* see if this is the log file */
|
|
if (is_not_the_log_file (hndl->fname) == 0)
|
|
{
|
|
/* If we are trying to map the log file, write to stderr, not to the experiment */
|
|
char errbuf[4096];
|
|
CALL_UTIL (snprintf)(errbuf, sizeof (errbuf),
|
|
"create_handle: can't mmap MAP_ANON (for %s): %s\n", hndl->fname, CALL_UTIL (strerror)(errno));
|
|
CALL_UTIL (write)(2, errbuf, CALL_UTIL (strlen)(errbuf));
|
|
}
|
|
else /* write the error message into the experiment */
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">MAP_ANON (for %s); create_handle</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEMAP, errno, hndl->fname);
|
|
return NULL;
|
|
}
|
|
TprintfT (DBG_LT2, " create_handle IO_TXT data buffer length=%ld (0x%lx) file='%s' memory=%p -- %p\n",
|
|
(long) (NBUFS * blksz), (long) (NBUFS * blksz), hndl->fname,
|
|
memory, memory + (NBUFS * blksz) - 1);
|
|
|
|
/* set up an array of buffers, pointing them to the virtual addresses */
|
|
TprintfT (DBG_LT2, "create_handle IO_TXT Buffer structures fname = `%s', NBUFS= %d, size = %ld (0x%lx)\n", fname,
|
|
NBUFS, (long) NBUFS * sizeof (Buffer), (long) NBUFS * sizeof (Buffer));
|
|
hndl->buffers = (Buffer*) __collector_allocCSize (__collector_heap, NBUFS * sizeof (Buffer), 1);
|
|
if (hndl->buffers == NULL)
|
|
{
|
|
TprintfT (0, "create_handle allocCSize for hndl->buffers failed\n");
|
|
CALL_UTIL (munmap)(memory, NBUFS * blksz);
|
|
return NULL;
|
|
}
|
|
for (int i = 0; i < NBUFS; i++)
|
|
{
|
|
Buffer *buf = &hndl->buffers[i];
|
|
buf->vaddr = memory + i * blksz;
|
|
buf->state = ST_FREE;
|
|
}
|
|
/* set the file pointer to the beginning of the file */
|
|
hndl->curpos = CUR_MAKE (0, 0, 0);
|
|
}
|
|
else
|
|
{
|
|
if (hndl->iotype == IO_BLK)
|
|
{
|
|
long nflow = CALL_UTIL (sysconf)(_SC_NPROCESSORS_ONLN);
|
|
if (nflow < 16)
|
|
nflow = 16;
|
|
hndl->nflow = (uint32_t) nflow;
|
|
}
|
|
else if (hndl->iotype == IO_SEQ)
|
|
hndl->nflow = 1;
|
|
TprintfT (DBG_LT2, "create_handle calling allocCSize blkstate fname=`%s' nflow=%d NCHUNKS=%d size=%ld (0x%lx)\n",
|
|
fname, hndl->nflow, NCHUNKS,
|
|
(long) (hndl->nflow * NCHUNKS * sizeof (uint32_t)),
|
|
(long) (hndl->nflow * NCHUNKS * sizeof (uint32_t)));
|
|
uint32_t *blkstate = (uint32_t*) __collector_allocCSize (__collector_heap, hndl->nflow * NCHUNKS * sizeof (uint32_t), 1);
|
|
if (blkstate == NULL)
|
|
return NULL;
|
|
for (int j = 0; j < hndl->nflow * NCHUNKS; ++j)
|
|
blkstate[j] = ST_INIT;
|
|
hndl->blkstate = blkstate;
|
|
TprintfT (DBG_LT2, "create_handle calling allocCSize blkoff fname=`%s' nflow=%d NCHUNKS=%d size=%ld (0x%lx)\n",
|
|
fname, hndl->nflow, NCHUNKS,
|
|
(long) (hndl->nflow * NCHUNKS * sizeof (uint32_t)),
|
|
(long) (hndl->nflow * NCHUNKS * sizeof (uint32_t)));
|
|
hndl->blkoff = (uint32_t*) __collector_allocCSize (__collector_heap, hndl->nflow * NCHUNKS * sizeof (uint32_t), 1);
|
|
if (hndl->blkoff == NULL)
|
|
return NULL;
|
|
hndl->nchnk = 0;
|
|
for (int j = 0; j < NCHUNKS; ++j)
|
|
{
|
|
hndl->chunks[j] = NULL;
|
|
hndl->chblk[j] = 0;
|
|
}
|
|
}
|
|
hndl->active = 1;
|
|
return hndl;
|
|
}
|
|
|
|
static void
|
|
deleteHandle (DataHandle *hndl)
|
|
{
|
|
if (hndl->active == 0)
|
|
return;
|
|
hndl->active = 0;
|
|
|
|
if (hndl->iotype == IO_BLK || hndl->iotype == IO_SEQ)
|
|
{
|
|
/* Delete all blocks. */
|
|
/* Since access to hndl->active is not synchronized it's still
|
|
* possible that we leave some blocks undeleted.
|
|
*/
|
|
for (int j = 0; j < hndl->nflow * NCHUNKS; ++j)
|
|
{
|
|
uint32_t oldstate = hndl->blkstate[j];
|
|
if (oldstate != ST_FREE)
|
|
continue;
|
|
/* Mark as busy */
|
|
uint32_t state = __collector_cas_32 (hndl->blkstate + j, oldstate, ST_BUSY);
|
|
if (state != oldstate)
|
|
continue;
|
|
deleteBlock (hndl, j / NCHUNKS, j % NCHUNKS);
|
|
}
|
|
}
|
|
else if (hndl->iotype == IO_TXT)
|
|
{
|
|
/*
|
|
* First, make sure that buffers are in some "coherent" state:
|
|
*
|
|
* At this point, the handle is no longer active. But some threads
|
|
* might already have passed the active-handle check and are now
|
|
* trying to schedule writes. So, set the handle pointer to "busy".
|
|
* This will prevent new writes from being scheduled. Threads that
|
|
* polling will time out.
|
|
*/
|
|
hrtime_t timeout = __collector_gethrtime () + 10 * ((hrtime_t) 1000000000);
|
|
volatile uint32_t busy = 0;
|
|
while (1)
|
|
{
|
|
uint32_t indx;
|
|
uint64_t opos, npos, foff;
|
|
int blk_off;
|
|
/* read the current pointer */
|
|
opos = hndl->curpos;
|
|
busy = CUR_BUSY (opos);
|
|
indx = CUR_INDX (opos);
|
|
foff = CUR_FOFF (opos);
|
|
if (busy == 1)
|
|
{
|
|
if (__collector_gethrtime () > timeout)
|
|
{
|
|
TprintfT (0, "deleteHandle ERROR: timeout cleaning up handle for %s\n", hndl->fname);
|
|
return;
|
|
}
|
|
continue;
|
|
}
|
|
blk_off = foff & (blksz - 1);
|
|
if (blk_off > 0)
|
|
foff += blksz - blk_off;
|
|
npos = CUR_MAKE (1, indx, foff);
|
|
|
|
/* try to update the handle position atomically */
|
|
if (__collector_cas_64p (&hndl->curpos, &opos, &npos) != opos)
|
|
continue;
|
|
|
|
/*
|
|
* If the last buffer won't be filled, account for
|
|
* the white space at the end so that the buffer will
|
|
* be deleted properly.
|
|
*/
|
|
if (blk_off > 0)
|
|
{
|
|
Buffer *buf = &hndl->buffers[indx];
|
|
if (__collector_subget_32 (&buf->left, blksz - blk_off) == 0)
|
|
deleteBuffer (buf);
|
|
}
|
|
break;
|
|
}
|
|
/* wait for buffers to be deleted */
|
|
timeout = __collector_gethrtime () + 10 * ((hrtime_t) 1000000000);
|
|
for (int i = 0; i < NBUFS; i++)
|
|
{
|
|
Buffer *buf = &hndl->buffers[i];
|
|
while (__collector_cas_32 (&buf->state, ST_FREE, ST_INIT) != ST_FREE)
|
|
{
|
|
if (__collector_gethrtime () > timeout)
|
|
{
|
|
TprintfT (0, "deleteHandle ERROR: timeout waiting for buffer %d for %s\n", i, hndl->fname);
|
|
return;
|
|
}
|
|
}
|
|
CALL_UTIL (munmap)(buf->vaddr, blksz);
|
|
}
|
|
|
|
/* free buffer array */
|
|
__collector_freeCSize (__collector_heap, hndl->buffers, NBUFS * sizeof (Buffer));
|
|
}
|
|
}
|
|
|
|
void
|
|
__collector_delete_handle (DataHandle *hndl)
|
|
{
|
|
if (hndl == NULL)
|
|
return;
|
|
deleteHandle (hndl);
|
|
}
|
|
|
|
static int
|
|
exp_size_ck (int nblocks, char *fname)
|
|
{
|
|
if (size_limit == 0)
|
|
return 0;
|
|
/* do an atomic add to the cur_size */
|
|
uint32_t old_size = cur_size;
|
|
uint32_t new_size;
|
|
for (;;)
|
|
{
|
|
new_size = __collector_cas_32 (&cur_size, old_size, old_size + nblocks);
|
|
if (new_size == old_size)
|
|
{
|
|
new_size = old_size + nblocks;
|
|
break;
|
|
}
|
|
old_size = new_size;
|
|
}
|
|
TprintfT (DBG_LT2, "exp_size_ck() adding %d block(s); new_size = %d, limit = %d blocks; fname = %s\n",
|
|
nblocks, new_size, size_limit, fname);
|
|
|
|
/* pause the entire collector if we have exceeded the limit */
|
|
if (old_size < size_limit && new_size >= size_limit)
|
|
{
|
|
TprintfT (0, "exp_size_ck() experiment size limit exceeded; new_size = %ld, limit = %ld blocks; fname = %s\n",
|
|
(long) new_size, (long) size_limit, fname);
|
|
(void) __collector_log_write ("<event kind=\"%s\" id=\"%d\">%ld blocks (each %ld bytes)</event>\n",
|
|
SP_JCMD_CWARN, COL_ERROR_SIZELIM, (long) size_limit, (long) blksz);
|
|
__collector_pause_m ("size-limit");
|
|
__collector_terminate_expt ();
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
__collector_set_size_limit (char *par)
|
|
{
|
|
if (!initialized)
|
|
init ();
|
|
|
|
int lim = CALL_UTIL (strtol)(par, &par, 0);
|
|
size_limit = (uint32_t) ((uint64_t) lim * 1024 * 1024 / blksz);
|
|
TprintfT (DBG_LT0, "collector_size_limit set to %d MB. = %d blocks\n",
|
|
lim, size_limit);
|
|
(void) __collector_log_write ("<setting limit=\"%d\"/>\n", lim);
|
|
return COL_ERROR_NONE;
|
|
}
|
|
|
|
/*
|
|
* IO_BLK and IO_SEQ files
|
|
*/
|
|
|
|
/*
|
|
* Allocate a chunk (nflow blocks) contiguously in virtual memory.
|
|
* Its blocks will be mmapped to the file individually.
|
|
*/
|
|
static int
|
|
allocateChunk (DataHandle *hndl, unsigned ichunk)
|
|
{
|
|
/*
|
|
* hndl->chunks[ichunk] is one of:
|
|
* - NULL (initial value)
|
|
* - CHUNK_BUSY (transition state when allocating the chunk)
|
|
* - some address (the allocated chunk)
|
|
*/
|
|
uint8_t *CHUNK_BUSY = (uint8_t *) 1;
|
|
hrtime_t timeout = 0;
|
|
while (1)
|
|
{
|
|
if (hndl->chunks[ichunk] > CHUNK_BUSY)
|
|
return 0; /* the chunk has already been allocated */
|
|
/* try to allocate the chunk (change: NULL => CHUNK_BUSY) */
|
|
if (__collector_cas_ptr (&hndl->chunks[ichunk], NULL, CHUNK_BUSY) == NULL)
|
|
{
|
|
/* allocate virtual memory */
|
|
uint8_t *newchunk = (uint8_t*) CALL_UTIL (mmap64)(0,
|
|
(size_t) (blksz * hndl->nflow),
|
|
PROT_READ | PROT_WRITE,
|
|
#if ARCH(SPARC)
|
|
MAP_SHARED | MAP_ANON,
|
|
#else
|
|
MAP_PRIVATE | MAP_ANON,
|
|
#endif
|
|
-1, (off64_t) 0);
|
|
if (newchunk == MAP_FAILED)
|
|
{
|
|
deleteHandle (hndl);
|
|
TprintfT (DBG_LT1, " allocateChunk mmap: start=0x%x length=%ld (0x%lx), offset=%d ret=%p\n",
|
|
0, (long) (blksz * hndl->nflow),
|
|
(long) (blksz * hndl->nflow), 0, newchunk);
|
|
TprintfT (0, "allocateChunk: can't mmap MAP_ANON (for %s): %s\n", hndl->fname, CALL_UTIL (strerror) (errno));
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">MAP_ANON (for %s)</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEMAP, errno, hndl->fname);
|
|
return 1;
|
|
}
|
|
|
|
/* assign allocated address to our chunk */
|
|
if (__collector_cas_ptr (&hndl->chunks[ichunk], CHUNK_BUSY, newchunk) != CHUNK_BUSY)
|
|
{
|
|
TprintfT (0, "allocateChunk: can't release chunk CAS lock for %s\n", hndl->fname);
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">couldn't release chunk CAS lock (%s)</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_GENERAL, hndl->fname);
|
|
}
|
|
__collector_inc_32 (&hndl->nchnk);
|
|
return 0;
|
|
}
|
|
|
|
/* check for time out */
|
|
if (timeout == 0)
|
|
timeout = __collector_gethrtime () + 10 * ((hrtime_t) 1000000000);
|
|
if (__collector_gethrtime () > timeout)
|
|
{
|
|
TprintfT (0, "allocateChunk: timeout for %s\n", hndl->fname);
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">timeout allocating chunk for %s</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_GENERAL, hndl->fname);
|
|
return 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Get the address for block (iflow,ichunk).
|
|
*/
|
|
static uint8_t *
|
|
getBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk)
|
|
{
|
|
return hndl->chunks[ichunk] + iflow * blksz;
|
|
}
|
|
|
|
/*
|
|
* Map block (iflow,ichunk) to the next part of the file.
|
|
*/
|
|
static int
|
|
remapBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk)
|
|
{
|
|
int rc = 0;
|
|
int fd;
|
|
/* Get the old file nblk and increment it atomically. */
|
|
uint32_t oldblk = hndl->nblk;
|
|
for (;;)
|
|
{
|
|
uint32_t newblk = __collector_cas_32 (&hndl->nblk, oldblk, oldblk + 1);
|
|
if (newblk == oldblk)
|
|
break;
|
|
oldblk = newblk;
|
|
}
|
|
off64_t offset = (off64_t) oldblk * blksz;
|
|
|
|
/* 6618470: disable thread cancellation */
|
|
int old_cstate;
|
|
pthread_setcancelstate (PTHREAD_CANCEL_DISABLE, &old_cstate);
|
|
|
|
/* Open the file. */
|
|
int iter = 0;
|
|
hrtime_t tso = __collector_gethrtime ();
|
|
for (;;)
|
|
{
|
|
fd = CALL_UTIL (open)(hndl->fname, O_RDWR, 0);
|
|
if (fd < 0)
|
|
{
|
|
if (errno == EMFILE)
|
|
{
|
|
/* too many open files */
|
|
iter++;
|
|
if (iter > 1000)
|
|
{
|
|
/* we've tried 1000 times; kick error back to caller */
|
|
char errmsg[MAXPATHLEN + 50];
|
|
hrtime_t teo = __collector_gethrtime ();
|
|
double deltato = (double) (teo - tso) / 1000000.;
|
|
(void) CALL_UTIL (snprintf) (errmsg, sizeof (errmsg), " t=%d, %s: open-retries-failed = %d, %3.6f ms.; remap",
|
|
__collector_thr_self (), hndl->fname, iter, deltato);
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">%s</event>\n",
|
|
SP_JCMD_COMMENT, COL_COMMENT_NONE, errmsg);
|
|
rc = 1;
|
|
goto exit;
|
|
}
|
|
/* keep trying */
|
|
continue;
|
|
}
|
|
deleteHandle (hndl);
|
|
TprintfT (0, "remapBlock: can't open file: %s: %s\n", hndl->fname, STR (CALL_UTIL (strerror)(errno)));
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">t=%llu, %s: remap </event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEOPN, errno,
|
|
(unsigned long long) __collector_thr_self (),
|
|
hndl->fname);
|
|
rc = 1;
|
|
goto exit;
|
|
}
|
|
else
|
|
break;
|
|
}
|
|
|
|
/* report number of retries of the open due to too many open fd's */
|
|
if (iter > 0)
|
|
{
|
|
char errmsg[MAXPATHLEN + 50];
|
|
hrtime_t teo = __collector_gethrtime ();
|
|
double deltato = (double) (teo - tso) / 1000000.;
|
|
(void) CALL_UTIL (snprintf) (errmsg, sizeof (errmsg), " t=%d, %s: open-retries = %d, %3.6f ms.; remap",
|
|
__collector_thr_self (), hndl->fname, iter, deltato);
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\">%s</event>\n",
|
|
SP_JCMD_COMMENT, COL_COMMENT_NONE, errmsg);
|
|
}
|
|
|
|
/* Ensure disk space is allocated and the block offset is 0 */
|
|
uint32_t zero = 0;
|
|
int n = CALL_UTIL (pwrite64)(fd, &zero, sizeof (zero), (off64_t) (offset + blksz - sizeof (zero)));
|
|
if (n <= 0)
|
|
{
|
|
deleteHandle (hndl);
|
|
TprintfT (0, "remapBlock: can't pwrite file: %s : errno=%d\n", hndl->fname, errno);
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">%s: remap</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_NOSPACE, errno, hndl->fname);
|
|
CALL_UTIL (close)(fd);
|
|
rc = 1;
|
|
goto exit;
|
|
}
|
|
hndl->blkoff[iflow * NCHUNKS + ichunk] = 0;
|
|
|
|
/* Map block to file */
|
|
uint8_t *bptr = getBlock (hndl, iflow, ichunk);
|
|
uint8_t *vaddr = (uint8_t *) CALL_UTIL (mmap64)(
|
|
(void*) bptr,
|
|
(size_t) blksz,
|
|
PROT_READ | PROT_WRITE,
|
|
MAP_SHARED | MAP_FIXED,
|
|
fd,
|
|
offset);
|
|
|
|
if (vaddr != bptr)
|
|
{
|
|
deleteHandle (hndl);
|
|
TprintfT (DBG_LT1, " remapBlock mmap: start=%p length=%ld (0x%lx) offset=0x%llx ret=%p\n",
|
|
bptr, (long) blksz, (long) blksz, (long long) offset, vaddr);
|
|
TprintfT (0, "remapBlock: can't mmap file: %s : errno=%d\n", hndl->fname, errno);
|
|
(void) __collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">%s: remap</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEMAP, errno, hndl->fname);
|
|
CALL_UTIL (close)(fd);
|
|
rc = 1;
|
|
goto exit;
|
|
}
|
|
CALL_UTIL (close)(fd);
|
|
|
|
if (hndl->exempt == 0)
|
|
exp_size_ck (1, hndl->fname);
|
|
else
|
|
Tprintf (DBG_LT1, "exp_size_ck() bypassed for %d block(s); exempt fname = %s\n",
|
|
1, hndl->fname);
|
|
exit:
|
|
/* Restore the previous cancellation state */
|
|
pthread_setcancelstate (old_cstate, NULL);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int
|
|
newBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk)
|
|
{
|
|
if (allocateChunk (hndl, ichunk) != 0)
|
|
return 1;
|
|
if (remapBlock (hndl, iflow, ichunk) != 0)
|
|
return 1;
|
|
|
|
/* Update the number of active blocks */
|
|
__collector_inc_32 (hndl->chblk + ichunk);
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
deleteBlock (DataHandle *hndl, unsigned iflow, unsigned ichunk)
|
|
{
|
|
uint8_t *bptr = getBlock (hndl, iflow, ichunk);
|
|
CALL_UTIL (munmap)((void*) bptr, blksz);
|
|
hndl->blkstate[iflow * NCHUNKS + ichunk] = ST_INIT;
|
|
|
|
/* Update the number of active blocks */
|
|
__collector_dec_32 (hndl->chblk + ichunk);
|
|
}
|
|
|
|
int
|
|
__collector_write_record (DataHandle *hndl, Common_packet *pckt)
|
|
{
|
|
if (hndl == NULL || !hndl->active)
|
|
return 1;
|
|
/* fill in the fields of the common packet structure */
|
|
if (pckt->type == 0)
|
|
pckt->type = hndl->kind;
|
|
if (pckt->tstamp == 0)
|
|
pckt->tstamp = __collector_gethrtime ();
|
|
if (pckt->lwp_id == 0)
|
|
pckt->lwp_id = __collector_lwp_self ();
|
|
if (pckt->thr_id == 0)
|
|
pckt->thr_id = __collector_thr_self ();
|
|
if (pckt->cpu_id == 0)
|
|
pckt->cpu_id = CALL_UTIL (getcpuid)();
|
|
if (pckt->tsize == 0)
|
|
pckt->tsize = sizeof (Common_packet);
|
|
TprintfT (DBG_LT3, "collector_write_record to %s, type:%d tsize:%d\n",
|
|
hndl->fname, pckt->type, pckt->tsize);
|
|
return __collector_write_packet (hndl, (CM_Packet*) pckt);
|
|
}
|
|
|
|
int
|
|
__collector_write_packet (DataHandle *hndl, CM_Packet *pckt)
|
|
{
|
|
if (hndl == NULL || !hndl->active)
|
|
return 1;
|
|
|
|
/* if the experiment is not open, there should be no writes */
|
|
if (__collector_expstate != EXP_OPEN)
|
|
{
|
|
#ifdef DEBUG
|
|
char *xstate;
|
|
switch (__collector_expstate)
|
|
{
|
|
case EXP_INIT:
|
|
xstate = "EXP_INIT";
|
|
break;
|
|
case EXP_OPEN:
|
|
xstate = "EXP_OPEN";
|
|
break;
|
|
case EXP_PAUSED:
|
|
xstate = "EXP_PAUSED";
|
|
break;
|
|
case EXP_CLOSED:
|
|
xstate = "EXP_CLOSED";
|
|
break;
|
|
default:
|
|
xstate = "Unknown";
|
|
break;
|
|
}
|
|
TprintfT (0, "collector_write_packet: write to %s while experiment state is %s\n",
|
|
hndl->fname, xstate);
|
|
#endif
|
|
return 1;
|
|
}
|
|
int recsz = pckt->tsize;
|
|
if (recsz > blksz)
|
|
{
|
|
TprintfT (0, "collector_write_packet: packet too long: %d (max %ld)\n", recsz, blksz);
|
|
return 1;
|
|
}
|
|
unsigned tid = (__collector_no_threads ? __collector_lwp_self () : __collector_thr_self ());
|
|
unsigned iflow = tid % hndl->nflow;
|
|
|
|
/* Acquire block */
|
|
uint32_t *sptr = &hndl->blkstate[iflow * NCHUNKS];
|
|
uint32_t state = ST_BUSY;
|
|
unsigned ichunk;
|
|
for (ichunk = 0; ichunk < NCHUNKS; ++ichunk)
|
|
{
|
|
uint32_t oldstate = sptr[ichunk];
|
|
if (oldstate == ST_BUSY)
|
|
continue;
|
|
/* Mark as busy */
|
|
state = __collector_cas_32 (sptr + ichunk, oldstate, ST_BUSY);
|
|
if (state == oldstate)
|
|
break;
|
|
if (state == ST_BUSY)
|
|
continue;
|
|
/* It's possible the state changed from ST_INIT to ST_FREE */
|
|
oldstate = state;
|
|
state = __collector_cas_32 (sptr + ichunk, oldstate, ST_BUSY);
|
|
if (state == oldstate)
|
|
break;
|
|
}
|
|
|
|
if (state == ST_BUSY || ichunk == NCHUNKS)
|
|
{
|
|
/* We are out of blocks for this data flow.
|
|
* We might switch to another flow but for now report and return.
|
|
*/
|
|
TprintfT (0, "collector_write_packet: all %d blocks on flow %d for %s are busy\n",
|
|
NCHUNKS, iflow, hndl->fname);
|
|
return 1;
|
|
}
|
|
|
|
if (state == ST_INIT && newBlock (hndl, iflow, ichunk) != 0)
|
|
return 1;
|
|
uint8_t *bptr = getBlock (hndl, iflow, ichunk);
|
|
uint32_t blkoff = hndl->blkoff[iflow * NCHUNKS + ichunk];
|
|
if (blkoff + recsz > blksz)
|
|
{
|
|
/* The record doesn't fit. Close the block */
|
|
if (blkoff < blksz)
|
|
{
|
|
Common_packet *closed = (Common_packet *) (bptr + blkoff);
|
|
closed->type = CLOSED_PCKT;
|
|
closed->tsize = blksz - blkoff; /* redundant */
|
|
}
|
|
if (remapBlock (hndl, iflow, ichunk) != 0)
|
|
return 1;
|
|
blkoff = hndl->blkoff[iflow * NCHUNKS + ichunk];
|
|
}
|
|
if (blkoff + recsz < blksz)
|
|
{
|
|
/* Set the empty padding */
|
|
Common_packet *empty = (Common_packet *) (bptr + blkoff + recsz);
|
|
empty->type = EMPTY_PCKT;
|
|
empty->tsize = blksz - blkoff - recsz;
|
|
}
|
|
__collector_memcpy (bptr + blkoff, pckt, recsz);
|
|
|
|
/* Release block */
|
|
if (hndl->active == 0)
|
|
{
|
|
deleteBlock (hndl, iflow, ichunk);
|
|
return 0;
|
|
}
|
|
hndl->blkoff[iflow * NCHUNKS + ichunk] += recsz;
|
|
sptr[ichunk] = ST_FREE;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* IO_TXT files
|
|
*
|
|
* IO_TXT covers the case where many threads are trying to write text messages
|
|
* sequentially (atomically) to a file. Examples include SP_LOG_FILE and SP_MAP_FILE.
|
|
*
|
|
* The file is not written directly, but by writing to mmapped virtual memory.
|
|
* The granularity of the mapping is a "Buffer". There may be as many as
|
|
* NBUFS buffers at any one time.
|
|
*
|
|
* The current position of the file is handled via hndl->curpos.
|
|
*
|
|
* * It is accessed atomically with 64-bit CAS instructions.
|
|
*
|
|
* * This 64-bit word encapsulates:
|
|
* - busy: a bit to lock access to hndl->curpos
|
|
* - indx: an index indicating which Buffer to use for the current position
|
|
* - foff: the file offset
|
|
*
|
|
* * The contents are accessed with:
|
|
* - unpack macros: CUR_BUSY CUR_INDX CUR_FOFF
|
|
* - pack macro : CUR_MAKE
|
|
*
|
|
* Conceptually, what happens when a thread wants to write a message is:
|
|
* - acquire the hndl->curpos "busy" lock
|
|
* . acquire and map new Buffers if needed to complete the message
|
|
* . update the file offset
|
|
* . release the lock
|
|
* - write to the corresponding buffers
|
|
*
|
|
* Each Buffer has a buf->left field that tracks how many more bytes
|
|
* need to be written to the Buffer. After a thread writes to a Buffer,
|
|
* it decrements buf->left atomically. When buf->left reaches 0, the
|
|
* Buffer (mapping) is deleted, freeing the Buffer for a new mapping.
|
|
*
|
|
* The actual implementation has some twists:
|
|
*
|
|
* * If the entire text message fits into the current Buffer -- that is,
|
|
* no new Buffers are needed -- the thread does not acquire the lock.
|
|
* It simply updates hndl->curpos atomically to the new file offset.
|
|
*
|
|
* * There are various timeouts to prevent hangs in case of abnormalities.
|
|
*/
|
|
static int
|
|
is_not_the_log_file (char *fname)
|
|
{
|
|
if (CALL_UTIL (strstr)(fname, SP_LOG_FILE) == NULL)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
mapBuffer (char *fname, Buffer *buf, off64_t foff)
|
|
{
|
|
int rc = 0;
|
|
/* open fname */
|
|
int fd = CALL_UTIL (open)(fname, O_RDWR, 0);
|
|
if (fd < 0)
|
|
{
|
|
TprintfT (0, "mapBuffer ERROR: can't open file: %s\n", fname);
|
|
if (is_not_the_log_file (fname))
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">%s: mapBuffer</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEOPN, errno, fname);
|
|
return 1;
|
|
}
|
|
TprintfT (DBG_LT2, "mapBuffer pwrite file %s at 0x%llx\n", fname, (long long) foff);
|
|
|
|
/* ensure disk space is allocated */
|
|
char nl = '\n';
|
|
int n = CALL_UTIL (pwrite64)(fd, &nl, sizeof (nl), (off64_t) (foff + blksz - sizeof (nl)));
|
|
if (n <= 0)
|
|
{
|
|
TprintfT (0, "mapBuffer ERROR: can't pwrite file %s at 0x%llx\n", fname,
|
|
(long long) (foff + blksz - sizeof (nl)));
|
|
if (is_not_the_log_file (fname))
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">%s: mapBuffer</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILETRNC, errno, fname);
|
|
rc = 1;
|
|
goto exit;
|
|
}
|
|
/* mmap buf->vaddr to fname at foff */
|
|
uint8_t *vaddr = CALL_UTIL (mmap64)(buf->vaddr, (size_t) blksz,
|
|
PROT_READ | PROT_WRITE, MAP_SHARED | MAP_FIXED, fd, foff);
|
|
if (vaddr != buf->vaddr)
|
|
{
|
|
TprintfT (DBG_LT1, " mapBuffer mmap: start=%p length=%ld (0x%lx) offset=0x%llx ret=%p\n",
|
|
buf->vaddr, blksz, blksz, (long long) foff, vaddr);
|
|
TprintfT (0, "mapBuffer ERROR: can't mmap %s: vaddr=%p size=%ld (0x%lx) ret=%p off=0x%llx errno=%d\n",
|
|
fname, buf->vaddr, blksz, blksz, vaddr, (long long) foff, errno);
|
|
if (is_not_the_log_file (fname))
|
|
__collector_log_write ("<event kind=\"%s\" id=\"%d\" ec=\"%d\">%s: mapBuffer</event>\n",
|
|
SP_JCMD_CERROR, COL_ERROR_FILEMAP, errno, fname);
|
|
rc = 1;
|
|
}
|
|
else
|
|
buf->left = blksz;
|
|
exit:
|
|
CALL_UTIL (close)(fd);
|
|
|
|
/* Should we check buffer size? Let's not since:
|
|
* - IO_TXT is typically not going to be that big
|
|
* - we want log.xml to be treated specially
|
|
*/
|
|
/* exp_size_ck( 1, fname ); */
|
|
return rc;
|
|
}
|
|
|
|
static int
|
|
newBuffer (DataHandle *hndl, uint64_t foff)
|
|
{
|
|
/* find a ST_FREE buffer and mark it ST_BUSY */
|
|
int ibuf;
|
|
for (ibuf = 0; ibuf < NBUFS; ibuf++)
|
|
if (__collector_cas_32 (&hndl->buffers[ibuf].state, ST_FREE, ST_BUSY) == ST_FREE)
|
|
break;
|
|
if (ibuf >= NBUFS)
|
|
{
|
|
TprintfT (0, "newBuffer ERROR: all buffers busy for %s\n", hndl->fname);
|
|
return -1;
|
|
}
|
|
Buffer *nbuf = hndl->buffers + ibuf;
|
|
|
|
/* map buffer */
|
|
if (mapBuffer (hndl->fname, nbuf, foff) != 0)
|
|
{
|
|
nbuf->state = ST_FREE;
|
|
ibuf = -1;
|
|
goto exit;
|
|
}
|
|
exit:
|
|
return ibuf;
|
|
}
|
|
|
|
static void
|
|
writeBuffer (Buffer *buf, int blk_off, char *src, int len)
|
|
{
|
|
__collector_memcpy (buf->vaddr + blk_off, src, len);
|
|
if (__collector_subget_32 (&buf->left, len) == 0)
|
|
deleteBuffer (buf);
|
|
}
|
|
|
|
static void
|
|
deleteBuffer (Buffer *buf)
|
|
{
|
|
buf->state = ST_FREE;
|
|
}
|
|
|
|
int
|
|
__collector_write_string (DataHandle *hndl, char *src, int len)
|
|
{
|
|
if (hndl == NULL || !hndl->active)
|
|
return 1;
|
|
if (len <= 0)
|
|
return 0;
|
|
|
|
hrtime_t timeout = __collector_gethrtime () + 20 * ((hrtime_t) 1000000000);
|
|
volatile uint32_t busy = 0;
|
|
while (1)
|
|
{
|
|
uint32_t indx;
|
|
uint64_t opos, foff, base;
|
|
int blk_off, buf_indices[NBUFS], ibuf, nbufs;
|
|
|
|
/* read and decode the current pointer */
|
|
opos = hndl->curpos;
|
|
busy = CUR_BUSY (opos);
|
|
indx = CUR_INDX (opos);
|
|
foff = CUR_FOFF (opos);
|
|
if (busy == 1)
|
|
{
|
|
if (__collector_gethrtime () > timeout)
|
|
{
|
|
/*
|
|
* E.g., if another thread deleted the handle
|
|
* after we checked hndl->active.
|
|
*/
|
|
TprintfT (0, "__collector_write_string ERROR: timeout writing length=%d to text file: %s\n", len, hndl->fname);
|
|
return 1;
|
|
}
|
|
continue;
|
|
}
|
|
|
|
/* initial block offset */
|
|
blk_off = foff & (blksz - 1);
|
|
|
|
/* number of new buffers to map */
|
|
int lastbuf = ((foff + len - 1) >> log2blksz); /* last block file index we will write */
|
|
int firstbuf = ((foff - 1) >> log2blksz); /* last block file index we have written */
|
|
nbufs = lastbuf - firstbuf;
|
|
TprintfT (DBG_LT2, "__collector_write_string firstbuf = %d, lastbuf = %d, nbufs = %d, log2blksz = %ld\n",
|
|
firstbuf, lastbuf, nbufs, log2blksz);
|
|
if (nbufs >= NBUFS)
|
|
{
|
|
Tprintf (0, "__collector_write_string ERROR: string of length %d too long to be written to text file: %s\n", len, hndl->fname);
|
|
return 1;
|
|
}
|
|
|
|
/* things are simple if we don't need new buffers */
|
|
if (nbufs == 0)
|
|
{
|
|
/* try to update the handle position atomically */
|
|
uint64_t npos = CUR_MAKE (0, indx, foff + len);
|
|
if (__collector_cas_64p (&hndl->curpos, &opos, &npos) != opos)
|
|
continue;
|
|
|
|
/* success! copy our string and we're done */
|
|
TprintfT (DBG_LT2, "__collector_write_string writeBuffer[%d]: vaddr = %p, len = %d, foff = %lld, '%s'\n",
|
|
indx, hndl->buffers[indx].vaddr, len, (long long) foff, src);
|
|
writeBuffer (&hndl->buffers[indx], foff & (blksz - 1), src, len);
|
|
break;
|
|
}
|
|
|
|
/* initialize the new signal mask */
|
|
sigset_t new_mask;
|
|
sigset_t old_mask;
|
|
CALL_UTIL (sigfillset)(&new_mask);
|
|
|
|
/* 6618470: disable thread cancellation */
|
|
int old_cstate;
|
|
pthread_setcancelstate (PTHREAD_CANCEL_DISABLE, &old_cstate);
|
|
/* block all signals */
|
|
CALL_UTIL (sigprocmask)(SIG_SETMASK, &new_mask, &old_mask);
|
|
|
|
/* but if we need new buffers, "lock" the handle pointer */
|
|
uint64_t lpos = CUR_MAKE (1, indx, foff);
|
|
if (__collector_cas_64p (&hndl->curpos, &opos, &lpos) != opos)
|
|
{
|
|
/* restore signal mask */
|
|
CALL_UTIL (sigprocmask)(SIG_SETMASK, &old_mask, NULL);
|
|
/* Restore the previous cancellation state */
|
|
pthread_setcancelstate (old_cstate, NULL);
|
|
continue;
|
|
}
|
|
|
|
/* map new buffers */
|
|
base = ((foff - 1) & ~(blksz - 1)); /* last buffer to have been mapped */
|
|
for (ibuf = 0; ibuf < nbufs; ibuf++)
|
|
{
|
|
base += blksz;
|
|
buf_indices[ibuf] = newBuffer (hndl, base);
|
|
if (buf_indices[ibuf] < 0)
|
|
break;
|
|
}
|
|
|
|
/* "unlock" the handle pointer */
|
|
uint64_t npos = CUR_MAKE (0, indx, foff);
|
|
if (ibuf == nbufs)
|
|
npos = CUR_MAKE (0, buf_indices[nbufs - 1], foff + len);
|
|
if (__collector_cas_64p (&hndl->curpos, &lpos, &npos) != lpos)
|
|
{
|
|
TprintfT (0, "__collector_write_string ERROR: file handle corrupted: %s\n", hndl->fname);
|
|
/*
|
|
* At this point, the handle is apparently corrupted and
|
|
* presumably locked. No telling what's going on. Still
|
|
* let's proceed and write our data and let a later thread
|
|
* raise an error if it encounters one.
|
|
*/
|
|
}
|
|
|
|
/* restore signal mask */
|
|
CALL_UTIL (sigprocmask)(SIG_SETMASK, &old_mask, NULL);
|
|
/* Restore the previous cancellation state */
|
|
pthread_setcancelstate (old_cstate, NULL);
|
|
|
|
/* if we couldn't map all the buffers we needed, don't write any part of the string */
|
|
if (ibuf < nbufs)
|
|
{
|
|
TprintfT (0, "__collector_write_string ERROR: can't map new buffer: %s\n", hndl->fname);
|
|
return 1;
|
|
}
|
|
|
|
/* write any data to the old block */
|
|
if (blk_off > 0)
|
|
{
|
|
TprintfT (DBG_LT2, "__collector_write_string partial writeBuffer[%d]: len=%ld, foff = %d '%s'\n",
|
|
indx, blksz - blk_off, blk_off, src);
|
|
writeBuffer (&hndl->buffers[indx], blk_off, src, blksz - blk_off);
|
|
src += blksz - blk_off;
|
|
len -= blksz - blk_off;
|
|
}
|
|
|
|
/* write data to the new blocks */
|
|
for (ibuf = 0; ibuf < nbufs; ibuf++)
|
|
{
|
|
int clen = blksz;
|
|
if (clen > len)
|
|
clen = len;
|
|
TprintfT (DBG_LT2, "__collector_write_string continue writeBuffer[%d]: len= %d, %s",
|
|
ibuf, clen, src);
|
|
writeBuffer (&hndl->buffers[buf_indices[ibuf]], 0, src, clen);
|
|
src += clen;
|
|
len -= clen;
|
|
}
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|