Home
last modified time | relevance | path

Searched refs:jobs (Results 1 – 25 of 305) sorted by relevance

12345678910>>...13

/external/icu/tools/srcgen/currysrc/libs/
Dorg.eclipse.core.jobs_3.7.0.v20150330-2103.jar ... .jobs org.eclipse.core.runtime.jobs.Job$1 extends org.eclipse.core ...
/external/eclipse-basebuilder/basebuilder-3.6.2/org.eclipse.releng.basebuilder/plugins/
Dorg.eclipse.core.jobs_3.5.1.R36x_v20100824.jar ... .jobs org.eclipse.core.internal.jobs.Deadlock extends java.lang.Object { private
/external/avahi/avahi-core/
Dprobe-sched.c49 AVAHI_LLIST_FIELDS(AvahiProbeJob, jobs);
56 AVAHI_LLIST_HEAD(AvahiProbeJob, jobs);
77 AVAHI_LLIST_PREPEND(AvahiProbeJob, jobs, s->history, pj); in job_new()
79 AVAHI_LLIST_PREPEND(AvahiProbeJob, jobs, s->jobs, pj); in job_new()
91 AVAHI_LLIST_REMOVE(AvahiProbeJob, jobs, s->history, pj); in job_free()
93 AVAHI_LLIST_REMOVE(AvahiProbeJob, jobs, s->jobs, pj); in job_free()
121 AVAHI_LLIST_REMOVE(AvahiProbeJob, jobs, s->jobs, pj); in job_mark_done()
122 AVAHI_LLIST_PREPEND(AvahiProbeJob, jobs, s->history, pj); in job_mark_done()
143 AVAHI_LLIST_HEAD_INIT(AvahiProbeJob, s->jobs); in avahi_probe_scheduler_new()
159 while (s->jobs) in avahi_probe_scheduler_clear()
[all …]
Dquery-sched.c58 AVAHI_LLIST_FIELDS(AvahiQueryJob, jobs);
74 AVAHI_LLIST_HEAD(AvahiQueryJob, jobs);
97 AVAHI_LLIST_PREPEND(AvahiQueryJob, jobs, s->history, qj); in job_new()
99 AVAHI_LLIST_PREPEND(AvahiQueryJob, jobs, s->jobs, qj); in job_new()
112 AVAHI_LLIST_REMOVE(AvahiQueryJob, jobs, s->history, qj); in job_free()
114 AVAHI_LLIST_REMOVE(AvahiQueryJob, jobs, s->jobs, qj); in job_free()
142 AVAHI_LLIST_REMOVE(AvahiQueryJob, jobs, s->jobs, qj); in job_mark_done()
143 AVAHI_LLIST_PREPEND(AvahiQueryJob, jobs, s->history, qj); in job_mark_done()
164 AVAHI_LLIST_HEAD_INIT(AvahiQueryJob, s->jobs); in avahi_query_scheduler_new()
182 while (s->jobs) in avahi_query_scheduler_clear()
[all …]
Dresponse-sched.c65 AVAHI_LLIST_FIELDS(AvahiResponseJob, jobs);
72 AVAHI_LLIST_HEAD(AvahiResponseJob, jobs);
95 AVAHI_LLIST_PREPEND(AvahiResponseJob, jobs, s->jobs, rj); in job_new()
97 AVAHI_LLIST_PREPEND(AvahiResponseJob, jobs, s->history, rj); in job_new()
99 AVAHI_LLIST_PREPEND(AvahiResponseJob, jobs, s->suppressed, rj); in job_new()
112 AVAHI_LLIST_REMOVE(AvahiResponseJob, jobs, s->jobs, rj); in job_free()
114 AVAHI_LLIST_REMOVE(AvahiResponseJob, jobs, s->history, rj); in job_free()
116 AVAHI_LLIST_REMOVE(AvahiResponseJob, jobs, s->suppressed, rj); in job_free()
144 AVAHI_LLIST_REMOVE(AvahiResponseJob, jobs, s->jobs, rj); in job_mark_done()
145 AVAHI_LLIST_PREPEND(AvahiResponseJob, jobs, s->history, rj); in job_mark_done()
[all …]
/external/chromium-trace/catapult/third_party/typ/typ/
Dpool.py23 def make_pool(host, jobs, callback, context, pre_fn, post_fn): argument
25 if jobs > 1:
26 return _ProcessPool(host, jobs, callback, context, pre_fn, post_fn)
28 return _AsyncPool(host, jobs, callback, context, pre_fn, post_fn)
60 def __init__(self, host, jobs, callback, context, pre_fn, post_fn): argument
62 self.jobs = jobs
69 for worker_num in range(1, jobs + 1):
180 def __init__(self, host, jobs, callback, context, pre_fn, post_fn): argument
182 self.jobs = jobs
/external/autotest/server/cros/dynamic_suite/
Djob_status.py87 def check_job_abort_status(afe, jobs): argument
101 for job in jobs:
115 def _abort_jobs_if_timedout(afe, jobs, start_time, timeout_mins): argument
131 for job in jobs:
138 def wait_for_jobs_to_start(afe, jobs, interval=DEFAULT_POLL_INTERVAL_SECONDS, argument
155 job_ids = [j.id for j in jobs]
157 if wait_timeout_mins and _abort_jobs_if_timedout(afe, jobs, start_time,
172 def wait_for_jobs_to_finish(afe, jobs, interval=DEFAULT_POLL_INTERVAL_SECONDS, argument
189 job_ids = [j.id for j in jobs]
191 if wait_timeout_mins and _abort_jobs_if_timedout(afe, jobs, start_time,
[all …]
Djob_status_unittest.py167 def expect_hosts_query_and_lock(self, jobs, manager, running_hosts, argument
183 for job in jobs:
429 jobs = [FakeJob(0, [FakeStatus('GOOD', 'T0', ''),
448 for status in jobs[4].statuses:
452 self.afe.get_jobs(id=jobs[0].id, finished=True).AndReturn([])
454 for job in jobs[1:]:
457 self.expect_result_gathering(jobs[0])
466 jobs)]
467 for job in jobs[:6]: # the 'GOOD' SERVER_JOB shouldn't be there.
475 jobs = [FakeJob(0, [FakeStatus('GOOD', 'T0', ''),
[all …]
/external/libdrm/tests/exynos/
Dexynos_fimg2d_event.c103 static void wait_all_jobs(struct g2d_job* jobs, unsigned num_jobs) in wait_all_jobs() argument
108 while (jobs[i].busy) in wait_all_jobs()
114 static struct g2d_job* free_job(struct g2d_job* jobs, unsigned num_jobs) in free_job() argument
119 if (jobs[i].busy == 0) in free_job()
120 return &jobs[i]; in free_job()
129 struct g2d_job *jobs = calloc(num_jobs, sizeof(struct g2d_job)); in g2d_work() local
135 jobs[i].id = i; in g2d_work()
143 j = free_job(jobs, num_jobs); in g2d_work()
182 wait_all_jobs(jobs, num_jobs); in g2d_work()
183 free(jobs); in g2d_work()
/external/autotest/scheduler/shard/
Dsimple_heartbeat_server.py78 jobs = models.Job.objects.filter(
82 return jobs[:job_limit] if job_limit is not None else jobs
95 def _create_packet(hosts, jobs): argument
98 'jobs': [j.serialize() for j in jobs]
124 job_time, jobs = self._get_jobs(board, job_limit)
126 serialize_time, heartbeat_packet = self._create_packet(hosts, jobs)
Dshard_client.py204 jobs = []
215 jobs.append(job_to_upload)
216 return jobs
226 def _get_hqes_for_jobs(self, jobs): argument
228 for job in jobs:
274 jobs = [job.serialize(include_dependencies=False) for job in job_objs]
275 logging.info('Uploading jobs %s', [j['id'] for j in jobs])
281 'jobs': jobs, 'hqes': hqes}
/external/autotest/scheduler/
Dhost_scheduler_unittests.py344 jobs = self.create_suite(num=2)
346 params=(jobs[0].id,))[0]
348 hqe = self.assign_host_to_job(host1, jobs[0], r)
349 self.verify_state(r, {jobs['parent_job'].id:1},
350 {host1.id: jobs['parent_job'].id})
353 self.verify_state(r, {jobs['parent_job'].id:1},
354 {host1.id: jobs['parent_job'].id})
356 self.assign_host_to_job(host2, jobs[1], r)
357 self.verify_state(r, {jobs['parent_job'].id:2},
358 {host1.id: jobs['parent_job'].id,
[all …]
/external/autotest/frontend/afe/
Dsite_rpc_interface_unittest.py513 jobs=upload_jobs, hqes=upload_hqes,
523 def _assert_shard_heartbeat_response(self, shard_hostname, retval, jobs=[], argument
529 (job.id, job.name, shard_hostname) for job in jobs]
549 self, jobs, hqes, shard_hostname='host1', argument
569 upload_jobs=jobs, upload_hqes=hqes)
575 upload_jobs=jobs, upload_hqes=hqes)
580 jobs, hqes = self._get_records_for_sending_to_master()
583 jobs=jobs, hqes=hqes, exception_to_throw=None)
592 jobs, hqes = self._get_records_for_sending_to_master()
595 jobs=jobs, hqes=hqes, exception_to_throw=None, aborted=True)
[all …]
Ddirect_afe_unittest.py14 jobs = afe.get_jobs()
15 self.assertEquals(len(jobs), 0)
25 jobs = afe.get_jobs()
26 self.assertEquals(len(jobs), 1)
/external/v8/tools/testrunner/objects/
Dpeer.py30 def __init__(self, address, jobs, rel_perf, pubkey): argument
32 self.jobs = jobs # integer: number of CPUs
44 (self.address, self.jobs, self.relative_performance,
74 return [self.address, self.jobs, self.relative_performance]
/external/glide/library/src/main/java/com/bumptech/glide/load/engine/
DEngine.java30 private final Map<Key, EngineJob> jobs; field in Engine
63 Map<Key, EngineJob> jobs, EngineKeyFactory keyFactory, in Engine() argument
79 if (jobs == null) { in Engine()
80 jobs = new HashMap<Key, EngineJob>(); in Engine()
82 this.jobs = jobs; in Engine()
177 EngineJob current = jobs.get(key); in load()
190 jobs.put(key, engineJob); in load()
237 jobs.remove(key); in onEngineJobComplete()
242 EngineJob current = jobs.get(key); in onEngineJobCancelled()
244 jobs.remove(key); in onEngineJobCancelled()
/external/v8/tools/testrunner/server/
Dpresence_handler.py60 jobs = data[1]
64 response = [STARTUP_RESPONSE, self.server.daemon.jobs,
69 p = peer.Peer(self.client_address[0], jobs, relative_perf,
75 jobs = data[1]
78 p = peer.Peer(self.client_address[0], jobs, perf, pubkey_fingerprint)
117 request = [STARTUP_REQUEST, self.daemon.jobs, self.daemon.relative_perf,
/external/chromium-trace/catapult/third_party/typ/typ/tests/
Dpool_test.py51 def run_basic_test(self, jobs): argument
54 pool = make_pool(host, jobs, _echo, context, _pre, _post)
65 expected_final_contexts = [expected_context for _ in range(jobs)]
159 jobs = 2
161 host, jobs, _stub, unpicklable_fn, None, None)
163 host, jobs, _stub, None, unpicklable_fn, None)
165 host, jobs, _stub, None, None, unpicklable_fn)
/external/llvm/utils/lit/lit/
Drun.py188 def execute_tests(self, display, jobs, max_time=None, argument
213 if jobs != 1 and use_processes and multiprocessing:
218 consumer = MultiprocessResultsConsumer(self, display, jobs)
232 provider = TestProvider(self.tests, jobs, queue_impl, canceled_flag)
249 if jobs == 1:
253 self._execute_tests_in_parallel(task_impl, provider, consumer, jobs)
264 def _execute_tests_in_parallel(self, task_impl, provider, consumer, jobs): argument
268 for i in range(jobs)]
/external/llvm/utils/
Dllvm-compilers-check255 def __init__(self, work_queue, jobs, argument
260 self.jobs = jobs
405 llvm=dict(debug=["-j" + str(self.jobs)],
406 release=["-j" + str(self.jobs)],
407 paranoid=["-j" + str(self.jobs)]),
408 dragonegg=dict(debug=["-j" + str(self.jobs)],
409 release=["-j" + str(self.jobs)],
410 paranoid=["-j" + str(self.jobs)]))
597 jobs = options.jobs // options.threads variable
598 if jobs == 0:
[all …]
/external/chromium-trace/catapult/perf_insights/perf_insights_build/
Dpi_report_to_html.py52 args.json, args.stop_on_error, args.jobs)
74 json_output=False, stop_on_error=False, jobs=1, quiet=False): argument
92 jobs, quiet)
107 jobs=1, quiet=False): argument
116 jobs=jobs)
/external/chromium-trace/catapult/telemetry/telemetry/testing/
Drun_tests.py133 runner.args.jobs = 1
136 runner.args.jobs = len(android_devs)
137 if runner.args.jobs == 0:
139 print 'Running tests with %d Android device(s).' % runner.args.jobs
143 runner.args.jobs = max(int(args.jobs) // 4, 1)
145 runner.args.jobs = max(int(args.jobs) // 2, 1)
/external/autotest/client/bin/self-test/tests/
D111-barrier-unordered-20x19 jobs = []
22 jobs.append([payload1, n])
24 job.parallel(*jobs)
D110-barrier-unordered-10x19 jobs = []
22 jobs.append([payload1, n])
24 job.parallel(*jobs)
D112-barrier-unordered-40x19 jobs = []
22 jobs.append([payload1, n])
24 job.parallel(*jobs)

12345678910>>...13