| #include <time.h> | 
 | #include <sys/time.h> | 
 |  | 
 | #include "fio.h" | 
 |  | 
 | static struct timeval genesis; | 
 | static unsigned long ns_granularity; | 
 |  | 
 | unsigned long utime_since(struct timeval *s, struct timeval *e) | 
 | { | 
 | 	long sec, usec, ret; | 
 |  | 
 | 	sec = e->tv_sec - s->tv_sec; | 
 | 	usec = e->tv_usec - s->tv_usec; | 
 | 	if (sec > 0 && usec < 0) { | 
 | 		sec--; | 
 | 		usec += 1000000; | 
 | 	} | 
 |  | 
 | 	sec *= 1000000UL; | 
 | 	ret = sec + usec; | 
 |  | 
 | 	/* | 
 | 	 * time warp bug on some kernels? | 
 | 	 */ | 
 | 	if (ret < 0) | 
 | 		ret = 0; | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | unsigned long utime_since_now(struct timeval *s) | 
 | { | 
 | 	struct timeval t; | 
 |  | 
 | 	fio_gettime(&t, NULL); | 
 | 	return utime_since(s, &t); | 
 | } | 
 |  | 
 | unsigned long mtime_since(struct timeval *s, struct timeval *e) | 
 | { | 
 | 	long sec, usec, ret; | 
 |  | 
 | 	sec = e->tv_sec - s->tv_sec; | 
 | 	usec = e->tv_usec - s->tv_usec; | 
 | 	if (sec > 0 && usec < 0) { | 
 | 		sec--; | 
 | 		usec += 1000000; | 
 | 	} | 
 |  | 
 | 	sec *= 1000UL; | 
 | 	usec /= 1000UL; | 
 | 	ret = sec + usec; | 
 |  | 
 | 	/* | 
 | 	 * time warp bug on some kernels? | 
 | 	 */ | 
 | 	if (ret < 0) | 
 | 		ret = 0; | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | unsigned long mtime_since_now(struct timeval *s) | 
 | { | 
 | 	struct timeval t; | 
 | 	void *p = __builtin_return_address(0); | 
 |  | 
 | 	fio_gettime(&t, p); | 
 | 	return mtime_since(s, &t); | 
 | } | 
 |  | 
 | unsigned long time_since_now(struct timeval *s) | 
 | { | 
 | 	return mtime_since_now(s) / 1000; | 
 | } | 
 |  | 
 | /* | 
 |  * busy looping version for the last few usec | 
 |  */ | 
 | void __usec_sleep(unsigned int usec) | 
 | { | 
 | 	struct timeval start; | 
 |  | 
 | 	fio_gettime(&start, NULL); | 
 | 	while (utime_since_now(&start) < usec) | 
 | 		nop; | 
 | } | 
 |  | 
 | void usec_sleep(struct thread_data *td, unsigned long usec) | 
 | { | 
 | 	struct timespec req; | 
 | 	struct timeval tv; | 
 |  | 
 | 	do { | 
 | 		unsigned long ts = usec; | 
 |  | 
 | 		if (usec < ns_granularity) { | 
 | 			__usec_sleep(usec); | 
 | 			break; | 
 | 		} | 
 |  | 
 | 		ts = usec - ns_granularity; | 
 |  | 
 | 		if (ts >= 1000000) { | 
 | 			req.tv_sec = ts / 1000000; | 
 | 			ts -= 1000000 * req.tv_sec; | 
 | 		} else | 
 | 			req.tv_sec = 0; | 
 |  | 
 | 		req.tv_nsec = ts * 1000; | 
 | 		fio_gettime(&tv, NULL); | 
 |  | 
 | 		if (nanosleep(&req, NULL) < 0) | 
 | 			break; | 
 |  | 
 | 		ts = utime_since_now(&tv); | 
 | 		if (ts >= usec) | 
 | 			break; | 
 |  | 
 | 		usec -= ts; | 
 | 	} while (!td->terminate); | 
 | } | 
 |  | 
 | void rate_throttle(struct thread_data *td, unsigned long time_spent, | 
 | 		   unsigned int bytes) | 
 | { | 
 | 	unsigned long usec_cycle; | 
 | 	unsigned int bs; | 
 |  | 
 | 	if (!td->o.rate && !td->o.rate_iops) | 
 | 		return; | 
 |  | 
 | 	if (td_rw(td)) | 
 | 		bs = td->o.rw_min_bs; | 
 | 	else if (td_read(td)) | 
 | 		bs = td->o.min_bs[DDIR_READ]; | 
 | 	else | 
 | 		bs = td->o.min_bs[DDIR_WRITE]; | 
 |  | 
 | 	usec_cycle = td->rate_usec_cycle * (bytes / bs); | 
 |  | 
 | 	if (time_spent < usec_cycle) { | 
 | 		unsigned long s = usec_cycle - time_spent; | 
 |  | 
 | 		td->rate_pending_usleep += s; | 
 |  | 
 | 		if (td->rate_pending_usleep >= 100000) { | 
 | 			struct timeval t; | 
 |  | 
 | 			fio_gettime(&t, NULL); | 
 | 			usec_sleep(td, td->rate_pending_usleep); | 
 | 			td->rate_pending_usleep -= utime_since_now(&t); | 
 | 		} | 
 | 	} else { | 
 | 		long overtime = time_spent - usec_cycle; | 
 |  | 
 | 		td->rate_pending_usleep -= overtime; | 
 | 	} | 
 | } | 
 |  | 
 | unsigned long mtime_since_genesis(void) | 
 | { | 
 | 	return mtime_since_now(&genesis); | 
 | } | 
 |  | 
 | static void fio_init time_init(void) | 
 | { | 
 | 	int i; | 
 |  | 
 | 	/* | 
 | 	 * Check the granularity of the nanosleep function | 
 | 	 */ | 
 | 	for (i = 0; i < 10; i++) { | 
 | 		struct timeval tv; | 
 | 		struct timespec ts; | 
 | 		unsigned long elapsed; | 
 |  | 
 | 		fio_gettime(&tv, NULL); | 
 | 		ts.tv_sec = 0; | 
 | 		ts.tv_nsec = 1000; | 
 |  | 
 | 		nanosleep(&ts, NULL); | 
 | 		elapsed = utime_since_now(&tv); | 
 |  | 
 | 		if (elapsed > ns_granularity) | 
 | 			ns_granularity = elapsed; | 
 | 	} | 
 | } | 
 |  | 
 | void set_genesis_time(void) | 
 | { | 
 | 	fio_gettime(&genesis, NULL); | 
 | } | 
 |  | 
 | void fill_start_time(struct timeval *t) | 
 | { | 
 | 	memcpy(t, &genesis, sizeof(genesis)); | 
 | } |